var/home/core/zuul-output/0000755000175000017500000000000015113561443014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113572346015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005160553315113572341017707 0ustar rootrootDec 02 13:04:28 crc systemd[1]: Starting Kubernetes Kubelet... Dec 02 13:04:28 crc restorecon[4685]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 13:04:28 crc restorecon[4685]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 02 13:04:29 crc kubenswrapper[4725]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 13:04:29 crc kubenswrapper[4725]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 02 13:04:29 crc kubenswrapper[4725]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 13:04:29 crc kubenswrapper[4725]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 13:04:29 crc kubenswrapper[4725]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 02 13:04:29 crc kubenswrapper[4725]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.145351 4725 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148362 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148391 4725 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148396 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148402 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148406 4725 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148413 4725 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148418 4725 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148421 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148425 4725 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148429 4725 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148433 4725 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148437 4725 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148441 4725 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148445 4725 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148450 4725 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148480 4725 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148486 4725 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148490 4725 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148495 4725 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148499 4725 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148528 4725 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148534 4725 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148538 4725 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148542 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148546 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148550 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148559 4725 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148568 4725 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148574 4725 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148580 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148585 4725 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148591 4725 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148595 4725 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148600 4725 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148605 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148610 4725 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148615 4725 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148620 4725 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148625 4725 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148630 4725 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148637 4725 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148643 4725 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148649 4725 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148654 4725 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148659 4725 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148666 4725 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148672 4725 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148676 4725 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148681 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148685 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148689 4725 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148693 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148697 4725 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148700 4725 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148704 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148710 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148714 4725 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148718 4725 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148722 4725 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148726 4725 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148730 4725 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148735 4725 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148738 4725 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148742 4725 feature_gate.go:330] unrecognized feature gate: Example Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148745 4725 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148749 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148752 4725 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148757 4725 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148760 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148764 4725 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.148767 4725 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148861 4725 flags.go:64] FLAG: --address="0.0.0.0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148870 4725 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148878 4725 flags.go:64] FLAG: --anonymous-auth="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148885 4725 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148891 4725 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148896 4725 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148903 4725 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148908 4725 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148912 4725 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148917 4725 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148921 4725 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148926 4725 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148930 4725 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148934 4725 flags.go:64] FLAG: --cgroup-root="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148939 4725 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148944 4725 flags.go:64] FLAG: --client-ca-file="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148948 4725 flags.go:64] FLAG: --cloud-config="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148953 4725 flags.go:64] FLAG: --cloud-provider="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148959 4725 flags.go:64] FLAG: --cluster-dns="[]" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148965 4725 flags.go:64] FLAG: --cluster-domain="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148969 4725 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148973 4725 flags.go:64] FLAG: --config-dir="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148977 4725 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148982 4725 flags.go:64] FLAG: --container-log-max-files="5" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148989 4725 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148994 4725 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.148999 4725 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149005 4725 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149010 4725 flags.go:64] FLAG: --contention-profiling="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149017 4725 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149023 4725 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149029 4725 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149035 4725 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149043 4725 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149049 4725 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149054 4725 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149059 4725 flags.go:64] FLAG: --enable-load-reader="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149064 4725 flags.go:64] FLAG: --enable-server="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149069 4725 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149077 4725 flags.go:64] FLAG: --event-burst="100" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149082 4725 flags.go:64] FLAG: --event-qps="50" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149087 4725 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149092 4725 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149096 4725 flags.go:64] FLAG: --eviction-hard="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149102 4725 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149107 4725 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149111 4725 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149115 4725 flags.go:64] FLAG: --eviction-soft="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149119 4725 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149123 4725 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149129 4725 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149133 4725 flags.go:64] FLAG: --experimental-mounter-path="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149137 4725 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149141 4725 flags.go:64] FLAG: --fail-swap-on="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149145 4725 flags.go:64] FLAG: --feature-gates="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149150 4725 flags.go:64] FLAG: --file-check-frequency="20s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149155 4725 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149159 4725 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149163 4725 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149168 4725 flags.go:64] FLAG: --healthz-port="10248" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149172 4725 flags.go:64] FLAG: --help="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149176 4725 flags.go:64] FLAG: --hostname-override="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149180 4725 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149185 4725 flags.go:64] FLAG: --http-check-frequency="20s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149189 4725 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149194 4725 flags.go:64] FLAG: --image-credential-provider-config="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149198 4725 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149202 4725 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149207 4725 flags.go:64] FLAG: --image-service-endpoint="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149211 4725 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149215 4725 flags.go:64] FLAG: --kube-api-burst="100" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149219 4725 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149224 4725 flags.go:64] FLAG: --kube-api-qps="50" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149228 4725 flags.go:64] FLAG: --kube-reserved="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149232 4725 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149236 4725 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149240 4725 flags.go:64] FLAG: --kubelet-cgroups="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149245 4725 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149250 4725 flags.go:64] FLAG: --lock-file="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149259 4725 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149267 4725 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149273 4725 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149284 4725 flags.go:64] FLAG: --log-json-split-stream="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149290 4725 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149295 4725 flags.go:64] FLAG: --log-text-split-stream="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149299 4725 flags.go:64] FLAG: --logging-format="text" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149304 4725 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149308 4725 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149312 4725 flags.go:64] FLAG: --manifest-url="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149317 4725 flags.go:64] FLAG: --manifest-url-header="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149323 4725 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149327 4725 flags.go:64] FLAG: --max-open-files="1000000" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149333 4725 flags.go:64] FLAG: --max-pods="110" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149337 4725 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149341 4725 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149345 4725 flags.go:64] FLAG: --memory-manager-policy="None" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149349 4725 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149354 4725 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149358 4725 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149362 4725 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149372 4725 flags.go:64] FLAG: --node-status-max-images="50" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149377 4725 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149381 4725 flags.go:64] FLAG: --oom-score-adj="-999" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149386 4725 flags.go:64] FLAG: --pod-cidr="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149391 4725 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149399 4725 flags.go:64] FLAG: --pod-manifest-path="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149403 4725 flags.go:64] FLAG: --pod-max-pids="-1" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149408 4725 flags.go:64] FLAG: --pods-per-core="0" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149412 4725 flags.go:64] FLAG: --port="10250" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149416 4725 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149421 4725 flags.go:64] FLAG: --provider-id="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149429 4725 flags.go:64] FLAG: --qos-reserved="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149440 4725 flags.go:64] FLAG: --read-only-port="10255" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149447 4725 flags.go:64] FLAG: --register-node="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149452 4725 flags.go:64] FLAG: --register-schedulable="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149479 4725 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149489 4725 flags.go:64] FLAG: --registry-burst="10" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149494 4725 flags.go:64] FLAG: --registry-qps="5" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149498 4725 flags.go:64] FLAG: --reserved-cpus="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149502 4725 flags.go:64] FLAG: --reserved-memory="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149508 4725 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149513 4725 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149517 4725 flags.go:64] FLAG: --rotate-certificates="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149521 4725 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149525 4725 flags.go:64] FLAG: --runonce="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149529 4725 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149533 4725 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149538 4725 flags.go:64] FLAG: --seccomp-default="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149543 4725 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149548 4725 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149561 4725 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149567 4725 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149572 4725 flags.go:64] FLAG: --storage-driver-password="root" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149577 4725 flags.go:64] FLAG: --storage-driver-secure="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149582 4725 flags.go:64] FLAG: --storage-driver-table="stats" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149587 4725 flags.go:64] FLAG: --storage-driver-user="root" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149592 4725 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149598 4725 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149603 4725 flags.go:64] FLAG: --system-cgroups="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149608 4725 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149618 4725 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149623 4725 flags.go:64] FLAG: --tls-cert-file="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149627 4725 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149634 4725 flags.go:64] FLAG: --tls-min-version="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149639 4725 flags.go:64] FLAG: --tls-private-key-file="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149643 4725 flags.go:64] FLAG: --topology-manager-policy="none" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149647 4725 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149652 4725 flags.go:64] FLAG: --topology-manager-scope="container" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149656 4725 flags.go:64] FLAG: --v="2" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149666 4725 flags.go:64] FLAG: --version="false" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149672 4725 flags.go:64] FLAG: --vmodule="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149677 4725 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.149681 4725 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149784 4725 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149789 4725 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149794 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149798 4725 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149801 4725 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149805 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149809 4725 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149812 4725 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149816 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149822 4725 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149825 4725 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149829 4725 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149832 4725 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149836 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149839 4725 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149843 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149846 4725 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149850 4725 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149853 4725 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149856 4725 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149860 4725 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149863 4725 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149867 4725 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149871 4725 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149875 4725 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149878 4725 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149882 4725 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149886 4725 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149889 4725 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149893 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149897 4725 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149900 4725 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149905 4725 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149910 4725 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149914 4725 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149918 4725 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149922 4725 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149926 4725 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149930 4725 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149934 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149939 4725 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149944 4725 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149948 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149952 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149956 4725 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149960 4725 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149972 4725 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149976 4725 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149980 4725 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149985 4725 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149989 4725 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149994 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.149999 4725 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150003 4725 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150008 4725 feature_gate.go:330] unrecognized feature gate: Example Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150012 4725 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150017 4725 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150022 4725 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150028 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150033 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150037 4725 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150041 4725 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150044 4725 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150048 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150052 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150055 4725 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150059 4725 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150062 4725 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150066 4725 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150069 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.150073 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.150085 4725 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.157663 4725 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.157711 4725 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157811 4725 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157824 4725 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157829 4725 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157833 4725 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157838 4725 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157844 4725 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157848 4725 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157853 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157858 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157863 4725 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157868 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157873 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157877 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157881 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157885 4725 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157888 4725 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157892 4725 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157896 4725 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157900 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157903 4725 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157907 4725 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157911 4725 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157916 4725 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157920 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157925 4725 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157929 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157934 4725 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157938 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157942 4725 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157947 4725 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157951 4725 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157955 4725 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157960 4725 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157965 4725 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157978 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157984 4725 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157988 4725 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157993 4725 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.157997 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158001 4725 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158005 4725 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158009 4725 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158013 4725 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158017 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158020 4725 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158024 4725 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158028 4725 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158031 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158035 4725 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158041 4725 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158048 4725 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158052 4725 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158057 4725 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158062 4725 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158067 4725 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158072 4725 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158076 4725 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158080 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158084 4725 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158087 4725 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158091 4725 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158094 4725 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158098 4725 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158101 4725 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158105 4725 feature_gate.go:330] unrecognized feature gate: Example Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158108 4725 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158112 4725 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158116 4725 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158121 4725 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158131 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158141 4725 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.158148 4725 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158316 4725 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158325 4725 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158329 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158332 4725 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158336 4725 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158340 4725 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158343 4725 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158347 4725 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158350 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158354 4725 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158357 4725 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158361 4725 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158364 4725 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158368 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158371 4725 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158376 4725 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158380 4725 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158384 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158391 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158395 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158399 4725 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158403 4725 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158407 4725 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158410 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158414 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158417 4725 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158421 4725 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158425 4725 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158430 4725 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158434 4725 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158438 4725 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158442 4725 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158446 4725 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158449 4725 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158479 4725 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158485 4725 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158489 4725 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158494 4725 feature_gate.go:330] unrecognized feature gate: Example Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158497 4725 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158501 4725 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158504 4725 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158508 4725 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158512 4725 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158515 4725 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158519 4725 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158523 4725 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158527 4725 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158531 4725 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158536 4725 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158541 4725 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158545 4725 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158549 4725 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158553 4725 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158557 4725 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158561 4725 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158565 4725 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158569 4725 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158573 4725 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158577 4725 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158580 4725 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158584 4725 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158588 4725 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158592 4725 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158597 4725 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158602 4725 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158606 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158610 4725 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158614 4725 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158618 4725 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158621 4725 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.158631 4725 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.158638 4725 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.158859 4725 server.go:940] "Client rotation is on, will bootstrap in background" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.162721 4725 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.162811 4725 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.163299 4725 server.go:997] "Starting client certificate rotation" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.163321 4725 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.163652 4725 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 02:49:09.787054843 +0000 UTC Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.163793 4725 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 613h44m40.623265532s for next certificate rotation Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.170045 4725 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.171515 4725 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.177993 4725 log.go:25] "Validated CRI v1 runtime API" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.189769 4725 log.go:25] "Validated CRI v1 image API" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.191045 4725 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.193550 4725 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-02-13-00-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.193582 4725 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.205882 4725 manager.go:217] Machine: {Timestamp:2025-12-02 13:04:29.204797947 +0000 UTC m=+0.161439662 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0121cc7c-04ba-4d57-99ad-07680ad4b3d9 BootID:2c6a5556-c77a-40a9-a29a-40e3b11740a7 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:6f:c8:d9 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:6f:c8:d9 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d6:4e:b4 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:2a:91:b7 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1a:2f:ab Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1c:9c:e5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:fe:79:1f:38:55:36 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:06:99:f4:3f:92:bf Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.206105 4725 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.206206 4725 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.206766 4725 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.206941 4725 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.206967 4725 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.207169 4725 topology_manager.go:138] "Creating topology manager with none policy" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.207182 4725 container_manager_linux.go:303] "Creating device plugin manager" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.207306 4725 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.207340 4725 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.207483 4725 state_mem.go:36] "Initialized new in-memory state store" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.207585 4725 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.208181 4725 kubelet.go:418] "Attempting to sync node with API server" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.208202 4725 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.208282 4725 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.208295 4725 kubelet.go:324] "Adding apiserver pod source" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.208307 4725 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.211621 4725 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.212054 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.212082 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.212870 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.212878 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.212884 4725 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.213709 4725 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214239 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214261 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214270 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214277 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214289 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214296 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214303 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214316 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214329 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214339 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214351 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214360 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.214702 4725 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.215455 4725 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.215745 4725 server.go:1280] "Started kubelet" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.216241 4725 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.216211 4725 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.216754 4725 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 02 13:04:29 crc systemd[1]: Started Kubernetes Kubelet. Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.217504 4725 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d67b80e99b290 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 13:04:29.2156996 +0000 UTC m=+0.172341295,LastTimestamp:2025-12-02 13:04:29.2156996 +0000 UTC m=+0.172341295,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.218180 4725 server.go:460] "Adding debug handlers to kubelet server" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.219706 4725 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.219752 4725 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.220189 4725 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.220956 4725 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.221105 4725 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.220749 4725 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.220130 4725 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 21:34:55.668010482 +0000 UTC Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.221134 4725 factory.go:55] Registering systemd factory Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.221159 4725 factory.go:221] Registration of the systemd container factory successfully Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.221949 4725 factory.go:153] Registering CRI-O factory Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.221965 4725 factory.go:221] Registration of the crio container factory successfully Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.222023 4725 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.222042 4725 factory.go:103] Registering Raw factory Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.222058 4725 manager.go:1196] Started watching for new ooms in manager Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.222721 4725 manager.go:319] Starting recovery of all containers Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.223473 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.223546 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.223707 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="200ms" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233352 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233411 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233434 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233448 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233509 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233525 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233539 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233558 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233574 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233591 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233605 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233622 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233657 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233678 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233690 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233704 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233722 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233736 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.233754 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234416 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234451 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234482 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234506 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234522 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234537 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234551 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234573 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234588 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234602 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.234617 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235210 4725 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235242 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235261 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235276 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235290 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235304 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235318 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235331 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235345 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235358 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235400 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235414 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235426 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235439 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235452 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235483 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235497 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235510 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235525 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235537 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235549 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235563 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235576 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235595 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235610 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235624 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235638 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235651 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235667 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235680 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235693 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235706 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235719 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235734 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235753 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235768 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235782 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235875 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235908 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235922 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235936 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235950 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235964 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.235978 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236016 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236033 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236048 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236064 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236079 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236094 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236108 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236123 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236136 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236151 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236164 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236178 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236192 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236207 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236261 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236277 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236293 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236307 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236324 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236340 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236356 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236373 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236389 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236404 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236419 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236436 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236450 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236483 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236500 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236517 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236533 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236553 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236572 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236588 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236603 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236618 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236633 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236646 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236662 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236675 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236695 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236708 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236720 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236737 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236748 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236760 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236774 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236787 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236801 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236814 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236827 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236839 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236852 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236868 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236884 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236897 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236910 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236923 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236936 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236950 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236963 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236977 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.236990 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237004 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237023 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237036 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237048 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237064 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237076 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237090 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237105 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237118 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237135 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237148 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237160 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237172 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237185 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237197 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237210 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237223 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237237 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237250 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237265 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237276 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237289 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237304 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237317 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237330 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237352 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237366 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237379 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237393 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237406 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237420 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237433 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237446 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237477 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237490 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237502 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237515 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237526 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237539 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237551 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237565 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237578 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237590 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237602 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237615 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237629 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237642 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237654 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237666 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237680 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237691 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237704 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237717 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237730 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237745 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237757 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237769 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237782 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237795 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237808 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237821 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237833 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237845 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237861 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237874 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237886 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237898 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237911 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237924 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237937 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237952 4725 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237964 4725 reconstruct.go:97] "Volume reconstruction finished" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.237974 4725 reconciler.go:26] "Reconciler: start to sync state" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.245787 4725 manager.go:324] Recovery completed Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.255503 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.259178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.259227 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.259238 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.261909 4725 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.261925 4725 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.261945 4725 state_mem.go:36] "Initialized new in-memory state store" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.265292 4725 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.266856 4725 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.266895 4725 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.266927 4725 kubelet.go:2335] "Starting kubelet main sync loop" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.266969 4725 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.267945 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.268009 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.269643 4725 policy_none.go:49] "None policy: Start" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.270585 4725 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.270617 4725 state_mem.go:35] "Initializing new in-memory state store" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.322141 4725 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.359655 4725 manager.go:334] "Starting Device Plugin manager" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.359721 4725 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.359736 4725 server.go:79] "Starting device plugin registration server" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.360245 4725 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.360262 4725 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.360645 4725 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.360753 4725 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.360760 4725 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.366583 4725 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.367675 4725 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.367741 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.368534 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.368564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.368572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.368676 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.368916 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.368954 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.369963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.370003 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.370016 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.370148 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.370697 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.370774 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372766 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372795 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372796 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372834 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372844 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372791 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372922 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372938 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.372808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.373054 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.373308 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.373420 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374337 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374365 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374535 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374558 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374566 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374653 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374756 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.374806 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.376374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.376398 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.376399 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.376422 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.376434 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.376406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.377261 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.377301 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.378029 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.378054 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.378072 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.425446 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="400ms" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440168 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440241 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440263 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440282 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440304 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440354 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440383 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440406 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440529 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440617 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440642 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440660 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440676 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440694 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.440737 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.460895 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.462728 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.462804 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.462818 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.462864 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.463611 4725 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542336 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542426 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542497 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542525 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542546 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542569 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542590 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542595 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542644 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542608 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542608 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542645 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542662 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542666 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542609 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542739 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542823 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542858 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542882 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542907 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542929 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542932 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542952 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542977 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.542985 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.543009 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.543010 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.543029 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.543054 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.543436 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.664057 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.665539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.665582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.665596 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.665624 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.666110 4725 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.707594 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.713572 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.728995 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-65c57cd12bbc0b145bacb833b34526b64bb2c099e53c53486471acebc9a33af8 WatchSource:0}: Error finding container 65c57cd12bbc0b145bacb833b34526b64bb2c099e53c53486471acebc9a33af8: Status 404 returned error can't find the container with id 65c57cd12bbc0b145bacb833b34526b64bb2c099e53c53486471acebc9a33af8 Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.730634 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7b64adc7b4391393c6c0791b3d4d1a476704bbeec9b965ee6b64e4ca597027bf WatchSource:0}: Error finding container 7b64adc7b4391393c6c0791b3d4d1a476704bbeec9b965ee6b64e4ca597027bf: Status 404 returned error can't find the container with id 7b64adc7b4391393c6c0791b3d4d1a476704bbeec9b965ee6b64e4ca597027bf Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.733645 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.741204 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: I1202 13:04:29.745984 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.782834 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-37f217116a9e08e617ce0bbc7fe1ab3096be4ddeb1ae7d326f3246ae6f4941fc WatchSource:0}: Error finding container 37f217116a9e08e617ce0bbc7fe1ab3096be4ddeb1ae7d326f3246ae6f4941fc: Status 404 returned error can't find the container with id 37f217116a9e08e617ce0bbc7fe1ab3096be4ddeb1ae7d326f3246ae6f4941fc Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.783659 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-28bfb4956dfcaa85ee9f0a1b9827cac50f43d1fe3705629ad0fc66d3b907d0e8 WatchSource:0}: Error finding container 28bfb4956dfcaa85ee9f0a1b9827cac50f43d1fe3705629ad0fc66d3b907d0e8: Status 404 returned error can't find the container with id 28bfb4956dfcaa85ee9f0a1b9827cac50f43d1fe3705629ad0fc66d3b907d0e8 Dec 02 13:04:29 crc kubenswrapper[4725]: W1202 13:04:29.788192 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ff9f1b26f551e769c8b0ad191c48d34fb1c3a0e3410e61246f4bffbd718bbb7b WatchSource:0}: Error finding container ff9f1b26f551e769c8b0ad191c48d34fb1c3a0e3410e61246f4bffbd718bbb7b: Status 404 returned error can't find the container with id ff9f1b26f551e769c8b0ad191c48d34fb1c3a0e3410e61246f4bffbd718bbb7b Dec 02 13:04:29 crc kubenswrapper[4725]: E1202 13:04:29.826860 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="800ms" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.066327 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.068562 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.068629 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.068641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.068671 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.069212 4725 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 02 13:04:30 crc kubenswrapper[4725]: W1202 13:04:30.088866 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.088984 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:30 crc kubenswrapper[4725]: W1202 13:04:30.137479 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.138007 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.216623 4725 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.221842 4725 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 04:42:11.212077626 +0000 UTC Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.221895 4725 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 471h37m40.990186673s for next certificate rotation Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.271314 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.271418 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"65c57cd12bbc0b145bacb833b34526b64bb2c099e53c53486471acebc9a33af8"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.273992 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.274061 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ff9f1b26f551e769c8b0ad191c48d34fb1c3a0e3410e61246f4bffbd718bbb7b"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.276208 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.276437 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"28bfb4956dfcaa85ee9f0a1b9827cac50f43d1fe3705629ad0fc66d3b907d0e8"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.277444 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.277500 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"37f217116a9e08e617ce0bbc7fe1ab3096be4ddeb1ae7d326f3246ae6f4941fc"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.277606 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.278748 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.278772 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.278781 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.279529 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.279564 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7b64adc7b4391393c6c0791b3d4d1a476704bbeec9b965ee6b64e4ca597027bf"} Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.279614 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.280600 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.280640 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.280650 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:30 crc kubenswrapper[4725]: W1202 13:04:30.534356 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.534440 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.628143 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="1.6s" Dec 02 13:04:30 crc kubenswrapper[4725]: W1202 13:04:30.634641 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.634727 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.869716 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.870712 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.870750 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.870760 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:30 crc kubenswrapper[4725]: I1202 13:04:30.870787 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:30 crc kubenswrapper[4725]: E1202 13:04:30.871305 4725 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.18:6443: connect: connection refused" node="crc" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.217224 4725 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.283370 4725 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a" exitCode=0 Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.283452 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.283492 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.284286 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.284313 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.284321 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.285013 4725 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2" exitCode=0 Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.285047 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.285122 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.291346 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.291384 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.291402 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.293189 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173" exitCode=0 Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.293261 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.293379 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.298654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.298702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.298714 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.302448 4725 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb" exitCode=0 Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.302557 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.302715 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.303895 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.303939 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.303952 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.305573 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.314645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.314676 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.314687 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.319237 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.319278 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.319292 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94"} Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.319336 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.320266 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.320300 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.320311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:31 crc kubenswrapper[4725]: W1202 13:04:31.757065 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.18:6443: connect: connection refused Dec 02 13:04:31 crc kubenswrapper[4725]: E1202 13:04:31.757158 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.18:6443: connect: connection refused" logger="UnhandledError" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.887659 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:31 crc kubenswrapper[4725]: I1202 13:04:31.892863 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.323660 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.323709 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.323723 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.324832 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.324947 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.325944 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.325961 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.325971 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.327833 4725 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d" exitCode=0 Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.327883 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.327966 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.338778 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.338811 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.338823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.342770 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343159 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343402 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343433 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343443 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343453 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0"} Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343725 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343745 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.343753 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.344114 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.344130 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.344137 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.471612 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.472843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.472877 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.472887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:32 crc kubenswrapper[4725]: I1202 13:04:32.472911 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.347665 4725 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda" exitCode=0 Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.347753 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.347766 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda"} Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.348850 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.348896 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.348909 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.352494 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab"} Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.352539 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a"} Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.352543 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.352590 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.352606 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.352590 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.353807 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.353843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.353859 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.354693 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.354719 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.354729 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.354845 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.354876 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:33 crc kubenswrapper[4725]: I1202 13:04:33.354887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.361704 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088"} Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.361784 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac"} Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.361798 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801"} Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.361818 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.361866 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.363297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.363351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:34 crc kubenswrapper[4725]: I1202 13:04:34.363361 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.368451 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f"} Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.368506 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090"} Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.368537 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.368626 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.369732 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.369774 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.369789 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.369934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.369972 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.369984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:35 crc kubenswrapper[4725]: I1202 13:04:35.606205 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.370355 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.370355 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.371307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.371345 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.371358 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.371878 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.371912 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.371924 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.484052 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.902678 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.903251 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.904836 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.904885 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:36 crc kubenswrapper[4725]: I1202 13:04:36.904903 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.373163 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.374522 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.374583 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.374604 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.720994 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.721179 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.722350 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.722388 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:37 crc kubenswrapper[4725]: I1202 13:04:37.722406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:38 crc kubenswrapper[4725]: I1202 13:04:38.042024 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:04:38 crc kubenswrapper[4725]: I1202 13:04:38.042259 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:38 crc kubenswrapper[4725]: I1202 13:04:38.043670 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:38 crc kubenswrapper[4725]: I1202 13:04:38.043704 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:38 crc kubenswrapper[4725]: I1202 13:04:38.043714 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:39 crc kubenswrapper[4725]: E1202 13:04:39.366700 4725 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 02 13:04:39 crc kubenswrapper[4725]: I1202 13:04:39.474169 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 02 13:04:39 crc kubenswrapper[4725]: I1202 13:04:39.474384 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:39 crc kubenswrapper[4725]: I1202 13:04:39.475738 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:39 crc kubenswrapper[4725]: I1202 13:04:39.475780 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:39 crc kubenswrapper[4725]: I1202 13:04:39.475792 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:40 crc kubenswrapper[4725]: I1202 13:04:40.602240 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:40 crc kubenswrapper[4725]: I1202 13:04:40.602391 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:40 crc kubenswrapper[4725]: I1202 13:04:40.603400 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:40 crc kubenswrapper[4725]: I1202 13:04:40.603431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:40 crc kubenswrapper[4725]: I1202 13:04:40.603442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:40 crc kubenswrapper[4725]: I1202 13:04:40.820484 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:41 crc kubenswrapper[4725]: I1202 13:04:41.381281 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:41 crc kubenswrapper[4725]: I1202 13:04:41.382245 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:41 crc kubenswrapper[4725]: I1202 13:04:41.382326 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:41 crc kubenswrapper[4725]: I1202 13:04:41.382356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.096205 4725 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.096817 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.217599 4725 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 02 13:04:42 crc kubenswrapper[4725]: E1202 13:04:42.230064 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 02 13:04:42 crc kubenswrapper[4725]: E1202 13:04:42.474534 4725 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 02 13:04:42 crc kubenswrapper[4725]: W1202 13:04:42.559135 4725 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.559434 4725 trace.go:236] Trace[629502538]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 13:04:32.557) (total time: 10001ms): Dec 02 13:04:42 crc kubenswrapper[4725]: Trace[629502538]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:04:42.559) Dec 02 13:04:42 crc kubenswrapper[4725]: Trace[629502538]: [10.00168261s] [10.00168261s] END Dec 02 13:04:42 crc kubenswrapper[4725]: E1202 13:04:42.559621 4725 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.711966 4725 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.712330 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.720162 4725 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 02 13:04:42 crc kubenswrapper[4725]: I1202 13:04:42.720488 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 02 13:04:43 crc kubenswrapper[4725]: I1202 13:04:43.004812 4725 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 13:04:43 crc kubenswrapper[4725]: I1202 13:04:43.004902 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 13:04:43 crc kubenswrapper[4725]: I1202 13:04:43.820816 4725 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 02 13:04:43 crc kubenswrapper[4725]: I1202 13:04:43.821593 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.612782 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.613037 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.613638 4725 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.613719 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.614569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.614777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.614904 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.618142 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.675433 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.677101 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.677160 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.677178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:45 crc kubenswrapper[4725]: I1202 13:04:45.677212 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:45 crc kubenswrapper[4725]: E1202 13:04:45.683403 4725 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.357319 4725 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.392882 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.393584 4725 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.393655 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.393863 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.393902 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:46 crc kubenswrapper[4725]: I1202 13:04:46.393914 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.714993 4725 trace.go:236] Trace[928041490]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 13:04:37.304) (total time: 10410ms): Dec 02 13:04:47 crc kubenswrapper[4725]: Trace[928041490]: ---"Objects listed" error: 10410ms (13:04:47.714) Dec 02 13:04:47 crc kubenswrapper[4725]: Trace[928041490]: [10.410183637s] [10.410183637s] END Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.715028 4725 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.715064 4725 trace.go:236] Trace[1811652018]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 13:04:32.976) (total time: 14738ms): Dec 02 13:04:47 crc kubenswrapper[4725]: Trace[1811652018]: ---"Objects listed" error: 14738ms (13:04:47.714) Dec 02 13:04:47 crc kubenswrapper[4725]: Trace[1811652018]: [14.738177371s] [14.738177371s] END Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.715097 4725 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.715637 4725 trace.go:236] Trace[1397923110]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 13:04:33.550) (total time: 14164ms): Dec 02 13:04:47 crc kubenswrapper[4725]: Trace[1397923110]: ---"Objects listed" error: 14164ms (13:04:47.715) Dec 02 13:04:47 crc kubenswrapper[4725]: Trace[1397923110]: [14.164818079s] [14.164818079s] END Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.715651 4725 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 02 13:04:47 crc kubenswrapper[4725]: I1202 13:04:47.715807 4725 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.218172 4725 apiserver.go:52] "Watching apiserver" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.221842 4725 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.222164 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-5zhgg","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.227769 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.227791 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.227803 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.228276 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.228212 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.228684 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.229551 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.229953 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.230216 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.230264 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.232390 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.233520 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.233548 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.233710 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.233731 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.234908 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.234915 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.235075 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.235975 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.236513 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.238968 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.238978 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.258641 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.269009 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.279515 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.289827 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.299856 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.310497 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.320118 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.322326 4725 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.331889 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.399214 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.400830 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab" exitCode=255 Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.400928 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab"} Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.411932 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419578 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419652 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419682 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419701 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419727 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419748 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419767 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419831 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419860 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419883 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419902 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419920 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419945 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419949 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.419985 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420011 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420049 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420073 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420096 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420123 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420144 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420155 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420166 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420190 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420210 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420230 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420251 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420314 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420334 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420336 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420407 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420431 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420454 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420496 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420516 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420536 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420561 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420581 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420604 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420622 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420645 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420664 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420683 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420700 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420716 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420723 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420734 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420788 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420811 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420855 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420873 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420876 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420909 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420929 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420946 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420962 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420982 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.420999 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421016 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421033 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421052 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421057 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421073 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421092 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421107 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421125 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421142 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421160 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421180 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421196 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421218 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421243 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.422136 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421686 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421714 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421721 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421719 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.421752 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.422075 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.422110 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.422839 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.422889 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423051 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423146 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423159 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423305 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423326 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423599 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423630 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423656 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423750 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423845 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424298 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424314 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424316 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424496 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424671 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424788 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.424955 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.425059 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.425093 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.423926 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.425625 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426010 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426150 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426537 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426612 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426749 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426764 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.426454 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.427050 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.427287 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.427531 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.427783 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.427944 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.429564 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.429654 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.422170 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.429909 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.429932 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.429957 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430005 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430038 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430068 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430102 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430136 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430171 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430200 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430231 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430263 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430289 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430322 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430354 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430384 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430419 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430786 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430899 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430939 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.431153 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.429929 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430378 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.431011 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.431089 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.431099 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.430950 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.431312 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:04:48.931288129 +0000 UTC m=+19.887929824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432509 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432618 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432643 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432655 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.431656 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432698 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432785 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.432842 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.431492 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.433290 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.433234 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.433573 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.433754 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434285 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434348 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434481 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434517 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434515 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434626 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434690 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434756 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434759 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434857 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.434893 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435115 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435318 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435376 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435413 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435445 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435492 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435640 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435676 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435710 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435742 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435775 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435806 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435836 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435871 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435905 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435934 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435968 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.435999 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436025 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436054 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436083 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436111 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436138 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436171 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436205 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436230 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436258 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436290 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436321 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436346 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436373 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436401 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436422 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436453 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436501 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436527 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436550 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436576 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436605 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.436629 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437186 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437277 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437288 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437374 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437420 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437294 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437525 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437682 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437732 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.437764 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.438274 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.438283 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.438445 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.438550 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.438679 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.438999 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439035 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439062 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439089 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439115 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439228 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439279 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439153 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439647 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439774 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.439304 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.440028 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.440064 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.440093 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.440861 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.440900 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441305 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441392 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441605 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441641 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441721 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441327 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441894 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.440308 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.442006 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.442056 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.441979 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.442306 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.443977 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.444065 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.444140 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.444425 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.445144 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.445180 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.445418 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.445444 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.445602 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.445750 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.446093 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.446184 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.446298 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.446429 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.446767 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.447254 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.447886 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.447986 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448074 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448264 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448287 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448613 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448686 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448930 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.448964 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449072 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449069 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449290 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449335 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449603 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449729 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.451826 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452085 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452124 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452152 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452173 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452196 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452217 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452238 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452295 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452320 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452339 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452359 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452381 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452401 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452420 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452492 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452515 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452536 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452554 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452581 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452599 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452617 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452636 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452655 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452686 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452705 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452724 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452744 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452765 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452867 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452916 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452938 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452958 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452977 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453000 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453020 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453039 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453058 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453075 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449651 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453161 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.449970 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453230 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453259 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453288 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/45c29b81-1929-415a-a671-606bae374a50-hosts-file\") pod \"node-resolver-5zhgg\" (UID: \"45c29b81-1929-415a-a671-606bae374a50\") " pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453493 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvjhc\" (UniqueName: \"kubernetes.io/projected/45c29b81-1929-415a-a671-606bae374a50-kube-api-access-bvjhc\") pod \"node-resolver-5zhgg\" (UID: \"45c29b81-1929-415a-a671-606bae374a50\") " pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453521 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453446 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453546 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453569 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453591 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453615 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453639 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453661 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453685 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453716 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453743 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453762 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453868 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453884 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455323 4725 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455354 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455379 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455390 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455408 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455422 4725 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.450293 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452025 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452488 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.452843 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.453786 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.454177 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.454271 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.454364 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.454416 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.454638 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455039 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455009 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455319 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.455399 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.456694 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:48.956661744 +0000 UTC m=+19.913303609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455935 4725 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455415 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455595 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455596 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455606 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455821 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455888 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.455957 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.456030 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457238 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457443 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457562 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457632 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457746 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457754 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457774 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.457825 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.457825 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.457916 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:48.957896174 +0000 UTC m=+19.914537869 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458034 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458142 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458303 4725 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458335 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458350 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458364 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458377 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458389 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458403 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458492 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458519 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458536 4725 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458555 4725 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458571 4725 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458587 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458606 4725 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458621 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458635 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458648 4725 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458662 4725 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458676 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458691 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458706 4725 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458686 4725 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458727 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458849 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458866 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458881 4725 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458895 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458909 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458923 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458939 4725 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458954 4725 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458969 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458983 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.458997 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459043 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459060 4725 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459074 4725 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459088 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459102 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459118 4725 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459132 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459147 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459161 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459510 4725 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459532 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459548 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459564 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459579 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459612 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459626 4725 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459640 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459653 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459666 4725 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459681 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459696 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459711 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459727 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459742 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459755 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459770 4725 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459785 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459801 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459816 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459831 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459845 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459859 4725 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459874 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459887 4725 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459900 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459912 4725 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459925 4725 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459938 4725 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459952 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459965 4725 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459977 4725 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.459991 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460005 4725 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460021 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460020 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460036 4725 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460052 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460068 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460081 4725 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460100 4725 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460113 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460127 4725 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460140 4725 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460152 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460166 4725 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460179 4725 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460191 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460202 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460214 4725 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460226 4725 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460237 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460249 4725 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460261 4725 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460275 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460287 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460299 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460311 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460324 4725 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460338 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460350 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460362 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460376 4725 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460387 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460400 4725 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460412 4725 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460425 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460437 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460449 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460477 4725 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460490 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460502 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460516 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460528 4725 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460540 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460553 4725 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.460567 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.462179 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.462179 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.462367 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.464221 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.464693 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.464862 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.465108 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.465552 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.468509 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.468933 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.469681 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.469708 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.469725 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.469787 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:48.969767122 +0000 UTC m=+19.926408817 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.471091 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.471473 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.471596 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.471838 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.472741 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.473246 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.473600 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.473620 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.473634 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.473690 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:48.973670586 +0000 UTC m=+19.930312471 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.474156 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.475747 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.479646 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.479719 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.479948 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480029 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480059 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480109 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480624 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480147 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480650 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.480936 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.481345 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.481428 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.481609 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.482632 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.482993 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.483507 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.483706 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.483944 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.486853 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.488343 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.489865 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.492924 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.493959 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.504105 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.504540 4725 scope.go:117] "RemoveContainer" containerID="e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.504850 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.515950 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.524059 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.536583 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.540606 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.543105 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.561853 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562043 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/45c29b81-1929-415a-a671-606bae374a50-hosts-file\") pod \"node-resolver-5zhgg\" (UID: \"45c29b81-1929-415a-a671-606bae374a50\") " pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562122 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvjhc\" (UniqueName: \"kubernetes.io/projected/45c29b81-1929-415a-a671-606bae374a50-kube-api-access-bvjhc\") pod \"node-resolver-5zhgg\" (UID: \"45c29b81-1929-415a-a671-606bae374a50\") " pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562196 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562280 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562350 4725 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562443 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562570 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562651 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562737 4725 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562836 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562895 4725 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.562953 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563009 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563062 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563114 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563173 4725 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563233 4725 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563294 4725 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563371 4725 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563445 4725 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563532 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563609 4725 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563694 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563781 4725 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563854 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563909 4725 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.563965 4725 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564032 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564090 4725 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564150 4725 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564210 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564268 4725 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564321 4725 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564414 4725 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564504 4725 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564593 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565522 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565633 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565762 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565844 4725 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565908 4725 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565973 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.566737 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.566841 4725 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.566897 4725 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.566984 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567107 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567215 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567328 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567426 4725 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567520 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567679 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567804 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.567947 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568074 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568136 4725 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568217 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568299 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568389 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568542 4725 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568653 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.568739 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.565479 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.564595 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.566102 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/45c29b81-1929-415a-a671-606bae374a50-hosts-file\") pod \"node-resolver-5zhgg\" (UID: \"45c29b81-1929-415a-a671-606bae374a50\") " pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569146 4725 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569432 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569576 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569675 4725 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569770 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569865 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.569969 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.570089 4725 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.581916 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xcgm9"] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.582850 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zc2vm"] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.583106 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8s8qq"] Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.583272 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.583382 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.583494 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.586211 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.586346 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.586515 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.587890 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.588259 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.588501 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.588724 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.588866 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvjhc\" (UniqueName: \"kubernetes.io/projected/45c29b81-1929-415a-a671-606bae374a50-kube-api-access-bvjhc\") pod \"node-resolver-5zhgg\" (UID: \"45c29b81-1929-415a-a671-606bae374a50\") " pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.588883 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.589110 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.588991 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.589471 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.589666 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.613739 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.640970 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.661834 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.670936 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-cni-binary-copy\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.670989 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-socket-dir-parent\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671015 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-k8s-cni-cncf-io\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671035 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-netns\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671088 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-cni-multus\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671179 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-etc-kubernetes\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671215 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fz4d\" (UniqueName: \"kubernetes.io/projected/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-kube-api-access-8fz4d\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671234 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cnibin\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671254 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-os-release\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671277 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c50af795-4a0f-45cd-b117-f6ddea79ee51-mcd-auth-proxy-config\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671301 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cni-binary-copy\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671317 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-hostroot\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671332 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-conf-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671349 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-multus-certs\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671368 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbpks\" (UniqueName: \"kubernetes.io/projected/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-kube-api-access-cbpks\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671383 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-system-cni-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671400 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-kubelet\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671422 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c50af795-4a0f-45cd-b117-f6ddea79ee51-rootfs\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671438 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c50af795-4a0f-45cd-b117-f6ddea79ee51-proxy-tls\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671496 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-daemon-config\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671528 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671545 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4gzp\" (UniqueName: \"kubernetes.io/projected/c50af795-4a0f-45cd-b117-f6ddea79ee51-kube-api-access-l4gzp\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671570 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-system-cni-dir\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671585 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-cni-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671603 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671617 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-cnibin\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671635 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-os-release\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.671649 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-cni-bin\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.686870 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.700792 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.714680 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.726680 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.736514 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.748202 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.760964 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772725 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-conf-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772780 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-multus-certs\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772806 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbpks\" (UniqueName: \"kubernetes.io/projected/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-kube-api-access-cbpks\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772829 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-system-cni-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772852 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-hostroot\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772876 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-kubelet\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772882 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-conf-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772691 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772957 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c50af795-4a0f-45cd-b117-f6ddea79ee51-rootfs\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.772898 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c50af795-4a0f-45cd-b117-f6ddea79ee51-rootfs\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773023 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-daemon-config\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773057 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c50af795-4a0f-45cd-b117-f6ddea79ee51-proxy-tls\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773087 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4gzp\" (UniqueName: \"kubernetes.io/projected/c50af795-4a0f-45cd-b117-f6ddea79ee51-kube-api-access-l4gzp\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773134 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773160 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-system-cni-dir\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773185 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-cni-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773200 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-multus-certs\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773212 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773258 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-cnibin\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773287 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-os-release\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773313 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-cni-bin\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773337 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-socket-dir-parent\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773359 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-k8s-cni-cncf-io\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773382 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-netns\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773403 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-cni-multus\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773436 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-cni-binary-copy\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773475 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-etc-kubernetes\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773506 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fz4d\" (UniqueName: \"kubernetes.io/projected/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-kube-api-access-8fz4d\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773524 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cnibin\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773547 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-os-release\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773578 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cni-binary-copy\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773597 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c50af795-4a0f-45cd-b117-f6ddea79ee51-mcd-auth-proxy-config\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.773989 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.774513 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c50af795-4a0f-45cd-b117-f6ddea79ee51-mcd-auth-proxy-config\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.774665 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-system-cni-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.774723 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-daemon-config\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.774758 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-system-cni-dir\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.774780 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775151 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-netns\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775206 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-hostroot\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775237 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-kubelet\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775280 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-cnibin\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775341 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-os-release\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775390 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-cni-bin\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775447 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-socket-dir-parent\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775534 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-run-k8s-cni-cncf-io\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775766 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-host-var-lib-cni-multus\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775916 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-os-release\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775973 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cnibin\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.776011 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-etc-kubernetes\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.775094 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-multus-cni-dir\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.776436 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-cni-binary-copy\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.776682 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-cni-binary-copy\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.779016 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c50af795-4a0f-45cd-b117-f6ddea79ee51-proxy-tls\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.789386 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbpks\" (UniqueName: \"kubernetes.io/projected/ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5-kube-api-access-cbpks\") pod \"multus-additional-cni-plugins-xcgm9\" (UID: \"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\") " pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.790615 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.803693 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4gzp\" (UniqueName: \"kubernetes.io/projected/c50af795-4a0f-45cd-b117-f6ddea79ee51-kube-api-access-l4gzp\") pod \"machine-config-daemon-zc2vm\" (UID: \"c50af795-4a0f-45cd-b117-f6ddea79ee51\") " pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.806602 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.813676 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fz4d\" (UniqueName: \"kubernetes.io/projected/4a8c02a6-36ad-4a9c-88b0-064dcc9b7327-kube-api-access-8fz4d\") pod \"multus-8s8qq\" (UID: \"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\") " pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.821695 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.831756 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.843880 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.848654 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.857751 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5zhgg" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.859720 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: W1202 13:04:48.863643 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c1ae719af7f12ffd3c106101b747636ab6500e246d637133fc6fd3b54681e3a8 WatchSource:0}: Error finding container c1ae719af7f12ffd3c106101b747636ab6500e246d637133fc6fd3b54681e3a8: Status 404 returned error can't find the container with id c1ae719af7f12ffd3c106101b747636ab6500e246d637133fc6fd3b54681e3a8 Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.864779 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.877952 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: W1202 13:04:48.878453 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45c29b81_1929_415a_a671_606bae374a50.slice/crio-f3c396ab18f38b21543fdb68c5860106e51dc50bd7b7a202d5be30cc45645097 WatchSource:0}: Error finding container f3c396ab18f38b21543fdb68c5860106e51dc50bd7b7a202d5be30cc45645097: Status 404 returned error can't find the container with id f3c396ab18f38b21543fdb68c5860106e51dc50bd7b7a202d5be30cc45645097 Dec 02 13:04:48 crc kubenswrapper[4725]: W1202 13:04:48.886248 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-04b1cf3a211caba6546181c1ca7ffa04f875dd390520c577c0ca0a7380f1115b WatchSource:0}: Error finding container 04b1cf3a211caba6546181c1ca7ffa04f875dd390520c577c0ca0a7380f1115b: Status 404 returned error can't find the container with id 04b1cf3a211caba6546181c1ca7ffa04f875dd390520c577c0ca0a7380f1115b Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.902527 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.911933 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.914337 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.916044 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8s8qq" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.934231 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.947954 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.947937 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.975530 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.975627 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.975659 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.975690 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:48 crc kubenswrapper[4725]: I1202 13:04:48.975725 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.975818 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.975885 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:49.975867169 +0000 UTC m=+20.932508864 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976276 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:04:49.976264249 +0000 UTC m=+20.932905954 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976384 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976407 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976422 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976475 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:49.976447443 +0000 UTC m=+20.933089138 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976538 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976577 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976594 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976624 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:49.976615027 +0000 UTC m=+20.933256722 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976685 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: E1202 13:04:48.976718 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:49.97670808 +0000 UTC m=+20.933349775 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:48 crc kubenswrapper[4725]: W1202 13:04:48.983154 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a8c02a6_36ad_4a9c_88b0_064dcc9b7327.slice/crio-6ece82a27c6566f65a927b3ed411a4d225fb3b617e6c424b496227a7ecafdbda WatchSource:0}: Error finding container 6ece82a27c6566f65a927b3ed411a4d225fb3b617e6c424b496227a7ecafdbda: Status 404 returned error can't find the container with id 6ece82a27c6566f65a927b3ed411a4d225fb3b617e6c424b496227a7ecafdbda Dec 02 13:04:49 crc kubenswrapper[4725]: W1202 13:04:49.008360 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab1ce6ac_3a16_49c5_b3db_6ca391e3aff5.slice/crio-12552bba06fd5e57de2f86d6034bb7bf7f24df6fbaca176c4a2eb47d547bdb1a WatchSource:0}: Error finding container 12552bba06fd5e57de2f86d6034bb7bf7f24df6fbaca176c4a2eb47d547bdb1a: Status 404 returned error can't find the container with id 12552bba06fd5e57de2f86d6034bb7bf7f24df6fbaca176c4a2eb47d547bdb1a Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.268055 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.268190 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.268432 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.268498 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.273008 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.273549 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.274856 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.275636 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.276669 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.277163 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.277758 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.278788 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.279427 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.280497 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.280703 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.280982 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.281995 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.283060 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.284255 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.284824 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.285801 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.286420 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.286889 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.287839 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.288522 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.288972 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.290064 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.290524 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.291569 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.292203 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.293305 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.294115 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.295224 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.295505 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.295903 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.296427 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.297753 4725 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.297857 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.300058 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.300680 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.301570 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.303215 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.303971 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.304910 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.305541 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.306681 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.307187 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.307164 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.308377 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.309138 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.310413 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.311115 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.312096 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.312654 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.313727 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.314211 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.315092 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.315622 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.316736 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.317614 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.318065 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.322732 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.339275 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.353490 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nnqsk"] Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.354372 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.356519 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.357193 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.357435 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.357539 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.357635 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.358379 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.361327 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.370809 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.383772 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.400133 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.405592 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.405636 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.405648 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4ec63971dfd2f2b59dd014481a7ce83d5857c3f4381aa311a638453cf0bf7632"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.408531 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.408588 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.408604 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"c941d1d08d3f42ab47355bb0f41635d993c6cec99fae1a0e5fc0808d4687ce20"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.413200 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerStarted","Data":"2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.413365 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerStarted","Data":"12552bba06fd5e57de2f86d6034bb7bf7f24df6fbaca176c4a2eb47d547bdb1a"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.414431 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.414478 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c1ae719af7f12ffd3c106101b747636ab6500e246d637133fc6fd3b54681e3a8"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.417707 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.419268 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.422356 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.422915 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.423987 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerStarted","Data":"7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.424025 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerStarted","Data":"6ece82a27c6566f65a927b3ed411a4d225fb3b617e6c424b496227a7ecafdbda"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.425110 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"04b1cf3a211caba6546181c1ca7ffa04f875dd390520c577c0ca0a7380f1115b"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.427361 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5zhgg" event={"ID":"45c29b81-1929-415a-a671-606bae374a50","Type":"ContainerStarted","Data":"1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.427403 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5zhgg" event={"ID":"45c29b81-1929-415a-a671-606bae374a50","Type":"ContainerStarted","Data":"f3c396ab18f38b21543fdb68c5860106e51dc50bd7b7a202d5be30cc45645097"} Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.434890 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.446554 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.461873 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.480441 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481061 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btq9p\" (UniqueName: \"kubernetes.io/projected/f939ed3f-9402-4a57-858f-0323084742a9-kube-api-access-btq9p\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481132 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-systemd\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481152 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-node-log\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481169 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481190 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-env-overrides\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481214 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481725 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481752 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-netns\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481822 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-systemd-units\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481840 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-etc-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481857 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-netd\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481902 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-slash\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481919 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-var-lib-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481936 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-ovn\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.481973 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-bin\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.482016 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-config\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.482031 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-script-lib\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.482067 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-log-socket\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.482105 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-kubelet\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.482129 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f939ed3f-9402-4a57-858f-0323084742a9-ovn-node-metrics-cert\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.506081 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.508176 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.519825 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.521978 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.531484 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.537855 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.553210 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.570797 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583268 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-kubelet\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583325 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f939ed3f-9402-4a57-858f-0323084742a9-ovn-node-metrics-cert\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583357 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btq9p\" (UniqueName: \"kubernetes.io/projected/f939ed3f-9402-4a57-858f-0323084742a9-kube-api-access-btq9p\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583401 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-systemd\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583424 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-node-log\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583413 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-kubelet\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583539 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583497 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-systemd\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583492 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583661 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-env-overrides\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583710 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583750 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583784 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-netns\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583807 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-systemd-units\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583835 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-etc-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583863 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-netd\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583935 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-slash\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.583967 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-var-lib-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584002 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-ovn\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584057 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-bin\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584095 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-config\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584131 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-script-lib\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584188 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-log-socket\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584287 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-log-socket\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584563 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-netd\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584657 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584705 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584743 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-netns\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584779 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-systemd-units\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584815 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-etc-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584862 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-ovn\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584903 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-slash\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.584950 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-var-lib-openvswitch\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.585299 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-env-overrides\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.585398 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-bin\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.585713 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-config\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.586133 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-script-lib\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.587368 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-node-log\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.592145 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.597386 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f939ed3f-9402-4a57-858f-0323084742a9-ovn-node-metrics-cert\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.609062 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btq9p\" (UniqueName: \"kubernetes.io/projected/f939ed3f-9402-4a57-858f-0323084742a9-kube-api-access-btq9p\") pod \"ovnkube-node-nnqsk\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.621419 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.644824 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.663391 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.680935 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.688000 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.727038 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.749977 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.771424 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.794728 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.810152 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.826472 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.845229 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.881397 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.909240 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.951848 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.986723 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.986826 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.986883 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.986913 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.986939 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:04:51.986909496 +0000 UTC m=+22.943551211 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.986984 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987040 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987082 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987078 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987175 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:51.987155572 +0000 UTC m=+22.943797267 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987172 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987198 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987211 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987253 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:51.987243614 +0000 UTC m=+22.943885409 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987095 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987321 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:51.987308406 +0000 UTC m=+22.943950101 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987602 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: E1202 13:04:49.987648 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:51.987639194 +0000 UTC m=+22.944280889 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:49 crc kubenswrapper[4725]: I1202 13:04:49.988097 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.025425 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.070750 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.267487 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:50 crc kubenswrapper[4725]: E1202 13:04:50.267669 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.432549 4725 generic.go:334] "Generic (PLEG): container finished" podID="ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5" containerID="2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56" exitCode=0 Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.432652 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerDied","Data":"2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56"} Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.434419 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2" exitCode=0 Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.434487 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2"} Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.434526 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"8d7ef5e525ac20549b36ea374fb347ef713cde9d74c4c74273f293bc146a06be"} Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.451947 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: E1202 13:04:50.453944 4725 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.467411 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.491071 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.514359 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.539618 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.562439 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.579028 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.599765 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.617568 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.632798 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.645481 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.665987 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.692230 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.708583 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.725811 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.747778 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.772769 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.812929 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.827202 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.831417 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.853889 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.865930 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.911939 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.955372 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:50 crc kubenswrapper[4725]: I1202 13:04:50.989305 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.029348 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.069649 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.107241 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.152198 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.188797 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.227771 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.242649 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-q658f"] Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.243264 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.268145 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:51 crc kubenswrapper[4725]: E1202 13:04:51.268271 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.268636 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:51 crc kubenswrapper[4725]: E1202 13:04:51.268702 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.270297 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.280939 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.300387 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.300402 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0803379-a77d-4799-a5ef-b2cfe94949a5-serviceca\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.300503 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvjcq\" (UniqueName: \"kubernetes.io/projected/e0803379-a77d-4799-a5ef-b2cfe94949a5-kube-api-access-cvjcq\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.300550 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0803379-a77d-4799-a5ef-b2cfe94949a5-host\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.320639 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.340909 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.390562 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.401803 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0803379-a77d-4799-a5ef-b2cfe94949a5-host\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.401880 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0803379-a77d-4799-a5ef-b2cfe94949a5-serviceca\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.401935 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvjcq\" (UniqueName: \"kubernetes.io/projected/e0803379-a77d-4799-a5ef-b2cfe94949a5-kube-api-access-cvjcq\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.401993 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0803379-a77d-4799-a5ef-b2cfe94949a5-host\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.403292 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e0803379-a77d-4799-a5ef-b2cfe94949a5-serviceca\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.439048 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvjcq\" (UniqueName: \"kubernetes.io/projected/e0803379-a77d-4799-a5ef-b2cfe94949a5-kube-api-access-cvjcq\") pod \"node-ca-q658f\" (UID: \"e0803379-a77d-4799-a5ef-b2cfe94949a5\") " pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.440102 4725 generic.go:334] "Generic (PLEG): container finished" podID="ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5" containerID="e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226" exitCode=0 Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.440212 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerDied","Data":"e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.445712 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.445768 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.445783 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.445795 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.445806 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.445818 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07"} Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.460667 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.490392 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.529413 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.557015 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q658f" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.572117 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: W1202 13:04:51.573785 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0803379_a77d_4799_a5ef_b2cfe94949a5.slice/crio-8ee17dfb41ac604af162df436a6a1d9793b563d0c56e2994f6f358fbe643acc2 WatchSource:0}: Error finding container 8ee17dfb41ac604af162df436a6a1d9793b563d0c56e2994f6f358fbe643acc2: Status 404 returned error can't find the container with id 8ee17dfb41ac604af162df436a6a1d9793b563d0c56e2994f6f358fbe643acc2 Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.608759 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.648555 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.691776 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.729387 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.771149 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.815772 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.848086 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.888058 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.927961 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:51 crc kubenswrapper[4725]: I1202 13:04:51.971005 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:51Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.008193 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.008389 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.008425 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008490 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:04:56.008427696 +0000 UTC m=+26.965069441 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.008570 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008586 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008608 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008624 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008692 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008718 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008732 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008750 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008699 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:56.008676162 +0000 UTC m=+26.965318007 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.008824 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008882 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:56.008866047 +0000 UTC m=+26.965507922 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.008900 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:56.008891777 +0000 UTC m=+26.965533482 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.009032 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.009148 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:04:56.009114633 +0000 UTC m=+26.965756328 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.015626 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.051219 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.084239 4725 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.088547 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.088591 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.088600 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.088731 4725 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.090029 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.140324 4725 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.140694 4725 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.141986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.142047 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.142060 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.142079 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.142092 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.155376 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.160261 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.160338 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.160354 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.160375 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.160388 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.166723 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.174356 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.180175 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.180222 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.180235 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.180253 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.180268 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.199016 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.204699 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.204755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.204769 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.204790 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.204804 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.223934 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.225941 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.230064 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.230109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.230119 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.230141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.230151 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.246370 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.246541 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.248244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.248304 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.248317 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.248340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.248353 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.252282 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.267257 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:52 crc kubenswrapper[4725]: E1202 13:04:52.267387 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.288592 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.327974 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.350790 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.350827 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.350835 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.350850 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.350859 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.367501 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.413545 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.461387 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.461485 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.461507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.461539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.461560 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.464732 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.466380 4725 generic.go:334] "Generic (PLEG): container finished" podID="ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5" containerID="945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276" exitCode=0 Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.466560 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerDied","Data":"945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.470056 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q658f" event={"ID":"e0803379-a77d-4799-a5ef-b2cfe94949a5","Type":"ContainerStarted","Data":"e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.470128 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q658f" event={"ID":"e0803379-a77d-4799-a5ef-b2cfe94949a5","Type":"ContainerStarted","Data":"8ee17dfb41ac604af162df436a6a1d9793b563d0c56e2994f6f358fbe643acc2"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.473273 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.491562 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.531673 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.563984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.564038 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.564051 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.564071 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.564086 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.571963 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.610337 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.650116 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.667362 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.667403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.667414 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.667431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.667441 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.690032 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.730970 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.769328 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.769380 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.769395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.769414 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.769426 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.775809 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.814219 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.849374 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.872224 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.872273 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.872289 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.872310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.872320 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.888711 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.927119 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.969838 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:52Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.974444 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.974494 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.974503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.974517 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:52 crc kubenswrapper[4725]: I1202 13:04:52.974527 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:52Z","lastTransitionTime":"2025-12-02T13:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.014441 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.049447 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.077080 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.077129 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.077141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.077159 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.077171 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.088930 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.133960 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.169195 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.180077 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.180122 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.180135 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.180153 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.180168 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.207018 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.249698 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.267580 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.267632 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:53 crc kubenswrapper[4725]: E1202 13:04:53.267733 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:53 crc kubenswrapper[4725]: E1202 13:04:53.267919 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.282631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.282683 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.282697 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.282715 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.282730 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.287842 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.328380 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.372938 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.385806 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.385843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.385852 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.385867 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.385875 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.405779 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.448416 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.481721 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.484262 4725 generic.go:334] "Generic (PLEG): container finished" podID="ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5" containerID="bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404" exitCode=0 Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.484355 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerDied","Data":"bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.487475 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.487584 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.487670 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.487755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.487917 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.491497 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.534805 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.568492 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.590087 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.590130 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.590139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.590155 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.590166 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.607872 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.648426 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.690515 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.692502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.692548 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.692563 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.692583 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.692596 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.728814 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.767577 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.795893 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.795931 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.795939 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.795956 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.795966 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.810990 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.865281 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.890416 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.901802 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.901851 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.901862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.901883 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.901911 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:53Z","lastTransitionTime":"2025-12-02T13:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.927562 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:53 crc kubenswrapper[4725]: I1202 13:04:53.969354 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.003771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.003799 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.003807 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.003821 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.003830 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.007540 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.048327 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.089902 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.105919 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.105953 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.105963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.105977 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.105990 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.128916 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.172571 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.208670 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.208815 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.208828 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.208846 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.208858 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.210779 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.253340 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.267419 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:54 crc kubenswrapper[4725]: E1202 13:04:54.267726 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.311749 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.311798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.311810 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.311829 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.311842 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.413603 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.413635 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.413647 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.413663 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.413675 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.491638 4725 generic.go:334] "Generic (PLEG): container finished" podID="ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5" containerID="a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44" exitCode=0 Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.491685 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerDied","Data":"a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.507273 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.516480 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.516512 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.516522 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.516539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.516550 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.521015 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.531963 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.548144 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.572551 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.587508 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.601911 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.613683 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.618540 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.618755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.618850 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.618945 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.619037 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.627910 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.647675 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.687393 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.721473 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.721512 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.721523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.721541 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.721553 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.727541 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.767502 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.813695 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.823247 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.823285 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.823297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.823313 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.823324 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.853304 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:54Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.925356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.925392 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.925401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.925415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:54 crc kubenswrapper[4725]: I1202 13:04:54.925424 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:54Z","lastTransitionTime":"2025-12-02T13:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.028062 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.028325 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.028400 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.028630 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.028802 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.131533 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.131840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.131942 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.132045 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.132124 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.234710 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.234754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.234763 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.234777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.234785 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.267773 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:55 crc kubenswrapper[4725]: E1202 13:04:55.267906 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.268255 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:55 crc kubenswrapper[4725]: E1202 13:04:55.268324 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.337435 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.337492 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.337503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.337523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.337537 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.439529 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.439713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.439808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.439892 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.439967 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.500945 4725 generic.go:334] "Generic (PLEG): container finished" podID="ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5" containerID="0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784" exitCode=0 Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.501048 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerDied","Data":"0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.519927 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.537742 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.544091 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.544126 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.544137 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.544156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.544168 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.554826 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.565334 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.577985 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.588759 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.602505 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.617602 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.631946 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.646508 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.646685 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.646707 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.646717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.646734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.646744 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.661257 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.679866 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.724771 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.744996 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.749941 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.749977 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.749990 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.750009 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.750023 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.773599 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:55Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.853109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.853175 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.853190 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.853212 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.853225 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.956109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.956148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.956160 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.956177 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:55 crc kubenswrapper[4725]: I1202 13:04:55.956186 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:55Z","lastTransitionTime":"2025-12-02T13:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.050641 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.050867 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.050956 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:05:04.050921623 +0000 UTC m=+35.007563388 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.051026 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.051062 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.051099 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051217 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051296 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:04.051272732 +0000 UTC m=+35.007914427 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051354 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051374 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051387 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051408 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051513 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051533 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051436 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:04.051425546 +0000 UTC m=+35.008067471 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051632 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:04.05160664 +0000 UTC m=+35.008248555 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051749 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.051787 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:04.051778034 +0000 UTC m=+35.008419949 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.058216 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.058250 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.058261 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.058275 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.058287 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.160708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.160744 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.160754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.160767 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.160776 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.262525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.262559 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.262568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.262580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.262591 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.267376 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:56 crc kubenswrapper[4725]: E1202 13:04:56.267602 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.364971 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.365008 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.365018 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.365033 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.365041 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.467431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.467508 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.467523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.467553 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.467568 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.507288 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" event={"ID":"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5","Type":"ContainerStarted","Data":"777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.512853 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.513107 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.513135 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.520774 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.533090 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.534105 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.535407 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.554065 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.566430 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.569672 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.569706 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.569718 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.569734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.569743 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.577927 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.591155 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.603235 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.613440 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.625595 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.638934 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.654446 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.665117 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.672103 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.672137 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.672146 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.672161 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.672173 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.676343 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.687656 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.697239 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.708001 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.725140 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.735559 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.745186 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.753639 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.763269 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.771480 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.773784 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.773813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.773822 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.773837 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.773845 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.782324 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.793531 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.808695 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.818262 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.830799 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.840765 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.852190 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.863964 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:56Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.875496 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.875537 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.875552 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.875569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.875580 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.977835 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.977882 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.977905 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.977924 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:56 crc kubenswrapper[4725]: I1202 13:04:56.977938 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:56Z","lastTransitionTime":"2025-12-02T13:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.080200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.080236 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.080248 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.080268 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.080278 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.182785 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.182819 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.182830 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.182845 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.182857 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.267290 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.267326 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:57 crc kubenswrapper[4725]: E1202 13:04:57.267479 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:57 crc kubenswrapper[4725]: E1202 13:04:57.267604 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.285860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.285896 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.285906 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.285921 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.285930 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.389162 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.389200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.389211 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.389230 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.389242 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.492311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.492367 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.492383 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.492404 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.492417 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.515852 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.600828 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.600887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.600906 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.600933 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.600949 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.704571 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.705137 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.705151 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.705176 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.705192 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.808505 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.808564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.808580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.808600 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.808614 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.911290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.911355 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.911374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.911403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:57 crc kubenswrapper[4725]: I1202 13:04:57.911426 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:57Z","lastTransitionTime":"2025-12-02T13:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.013789 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.013839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.013849 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.013865 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.013875 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.116523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.116568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.116578 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.116595 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.116608 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.218882 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.218919 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.218929 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.218943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.218953 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.267629 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:04:58 crc kubenswrapper[4725]: E1202 13:04:58.267755 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.321671 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.321710 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.321719 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.321736 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.321749 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.424503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.424544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.424553 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.424569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.424578 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.520544 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/0.log" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.523113 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7" exitCode=1 Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.523206 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.523971 4725 scope.go:117] "RemoveContainer" containerID="4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.526648 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.526730 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.526749 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.526766 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.526782 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.543209 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.582041 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.599273 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.611963 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.661484 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.663157 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.663196 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.663206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.663221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.663231 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.680530 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.691834 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.705094 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.715437 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.726806 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.740874 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.764403 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:57Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795699 6028 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795746 6028 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.796562 6028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 13:04:57.796603 6028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1202 13:04:57.796622 6028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1202 13:04:57.796649 6028 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 13:04:57.796677 6028 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 13:04:57.796650 6028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 13:04:57.796712 6028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 13:04:57.796734 6028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1202 13:04:57.796720 6028 factory.go:656] Stopping watch factory\\\\nI1202 13:04:57.796766 6028 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.766158 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.766217 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.766235 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.766265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.766283 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.781651 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.792939 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.803120 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:58Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.869070 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.869128 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.869139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.869201 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.869212 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.972110 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.972160 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.972170 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.972188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:58 crc kubenswrapper[4725]: I1202 13:04:58.972200 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:58Z","lastTransitionTime":"2025-12-02T13:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.077926 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.077997 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.078009 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.078033 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.078057 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.180502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.180556 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.180568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.180587 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.180602 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.267558 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:04:59 crc kubenswrapper[4725]: E1202 13:04:59.267683 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.267764 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:04:59 crc kubenswrapper[4725]: E1202 13:04:59.267888 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.283262 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.283309 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.283320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.283338 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.283351 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.285411 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.301973 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:57Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795699 6028 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795746 6028 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.796562 6028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 13:04:57.796603 6028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1202 13:04:57.796622 6028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1202 13:04:57.796649 6028 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 13:04:57.796677 6028 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 13:04:57.796650 6028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 13:04:57.796712 6028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 13:04:57.796734 6028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1202 13:04:57.796720 6028 factory.go:656] Stopping watch factory\\\\nI1202 13:04:57.796766 6028 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.312719 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.322122 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.332167 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.343058 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.354390 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.365533 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.384779 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.384818 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.384830 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.384848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.384860 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.389612 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.402893 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.413648 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.424127 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.439906 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.448509 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.459361 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.487342 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.487397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.487406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.487425 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.487437 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.527548 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/0.log" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.530398 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.530540 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.543287 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.553392 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.565622 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.588240 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.590226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.590276 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.590290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.590310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.590321 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.611087 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.622907 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.632235 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.644685 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.653479 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.663891 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.675240 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.688497 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.692144 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.692192 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.692202 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.692220 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.692314 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.700135 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.754044 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.777061 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:57Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795699 6028 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795746 6028 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.796562 6028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 13:04:57.796603 6028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1202 13:04:57.796622 6028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1202 13:04:57.796649 6028 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 13:04:57.796677 6028 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 13:04:57.796650 6028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 13:04:57.796712 6028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 13:04:57.796734 6028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1202 13:04:57.796720 6028 factory.go:656] Stopping watch factory\\\\nI1202 13:04:57.796766 6028 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.795958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.796009 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.796024 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.796045 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.796055 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.899411 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.899487 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.899499 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.899518 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:04:59 crc kubenswrapper[4725]: I1202 13:04:59.899530 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:04:59Z","lastTransitionTime":"2025-12-02T13:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.002304 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.002359 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.002372 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.002396 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.002410 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.105508 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.105565 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.105586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.105606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.105618 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.208429 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.208500 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.208511 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.208533 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.208543 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.267312 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:00 crc kubenswrapper[4725]: E1202 13:05:00.267973 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.310968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.311415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.311646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.311806 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.311995 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.415616 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.415693 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.415715 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.415754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.415774 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.518918 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.518982 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.518994 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.519027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.519043 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.536241 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/1.log" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.536759 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/0.log" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.541312 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea" exitCode=1 Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.541360 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.541440 4725 scope.go:117] "RemoveContainer" containerID="4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.542148 4725 scope.go:117] "RemoveContainer" containerID="73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea" Dec 02 13:05:00 crc kubenswrapper[4725]: E1202 13:05:00.542401 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.557230 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.568955 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.578371 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.591032 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.604068 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.618054 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.621780 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.621907 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.621978 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.622054 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.622145 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.631030 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.651748 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.669818 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:57Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795699 6028 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795746 6028 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.796562 6028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 13:04:57.796603 6028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1202 13:04:57.796622 6028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1202 13:04:57.796649 6028 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 13:04:57.796677 6028 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 13:04:57.796650 6028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 13:04:57.796712 6028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 13:04:57.796734 6028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1202 13:04:57.796720 6028 factory.go:656] Stopping watch factory\\\\nI1202 13:04:57.796766 6028 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.680265 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.692075 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.704515 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.722354 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.724425 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.724495 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.724509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.724532 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.724545 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.734994 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.748204 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:00Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.826643 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.826680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.826690 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.826707 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.826719 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.929177 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.929244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.929262 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.929287 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:00 crc kubenswrapper[4725]: I1202 13:05:00.929305 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:00Z","lastTransitionTime":"2025-12-02T13:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.032654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.032691 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.032699 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.032712 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.032721 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.136233 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.136298 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.136310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.136337 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.136351 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.239350 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.239394 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.239407 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.239433 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.239443 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.267338 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:01 crc kubenswrapper[4725]: E1202 13:05:01.267476 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.267441 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:01 crc kubenswrapper[4725]: E1202 13:05:01.267835 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.297311 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh"] Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.297965 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.300005 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.300009 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.312742 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.335812 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.341340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.341371 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.341384 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.341401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.341411 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.349375 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.361500 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.375876 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.388033 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.399396 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.402764 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e61071d0-15cb-43d9-a3f1-df802d821e1f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.402799 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwnlt\" (UniqueName: \"kubernetes.io/projected/e61071d0-15cb-43d9-a3f1-df802d821e1f-kube-api-access-pwnlt\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.402817 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e61071d0-15cb-43d9-a3f1-df802d821e1f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.402925 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e61071d0-15cb-43d9-a3f1-df802d821e1f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.409922 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.423156 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.433020 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.443393 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.443438 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.443447 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.443483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.443494 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.445681 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.459430 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.477832 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c77ae8404ea2e90049a9b4f14a0a9403d46338753a435858aa8b5125155f8f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:57Z\\\",\\\"message\\\":\\\"r.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795699 6028 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.795746 6028 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 13:04:57.796562 6028 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 13:04:57.796603 6028 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1202 13:04:57.796622 6028 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1202 13:04:57.796649 6028 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 13:04:57.796677 6028 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 13:04:57.796650 6028 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 13:04:57.796712 6028 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 13:04:57.796734 6028 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1202 13:04:57.796720 6028 factory.go:656] Stopping watch factory\\\\nI1202 13:04:57.796766 6028 ovnkube.go:599] Stopped ovnkube\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.503557 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e61071d0-15cb-43d9-a3f1-df802d821e1f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.503606 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwnlt\" (UniqueName: \"kubernetes.io/projected/e61071d0-15cb-43d9-a3f1-df802d821e1f-kube-api-access-pwnlt\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.503627 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e61071d0-15cb-43d9-a3f1-df802d821e1f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.503648 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e61071d0-15cb-43d9-a3f1-df802d821e1f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.504561 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.504923 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e61071d0-15cb-43d9-a3f1-df802d821e1f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.505214 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e61071d0-15cb-43d9-a3f1-df802d821e1f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.513142 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e61071d0-15cb-43d9-a3f1-df802d821e1f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.515921 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.520416 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwnlt\" (UniqueName: \"kubernetes.io/projected/e61071d0-15cb-43d9-a3f1-df802d821e1f-kube-api-access-pwnlt\") pod \"ovnkube-control-plane-749d76644c-tdvqh\" (UID: \"e61071d0-15cb-43d9-a3f1-df802d821e1f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.535919 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.545572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.545619 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.545637 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.545655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.545666 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.546521 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/1.log" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.550090 4725 scope.go:117] "RemoveContainer" containerID="73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea" Dec 02 13:05:01 crc kubenswrapper[4725]: E1202 13:05:01.550259 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.566924 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.579418 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.592634 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.612262 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.629776 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.654608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.654636 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.654645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.654660 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.654670 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.662119 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.687168 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.707195 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.732707 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.748384 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.757545 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.757605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.757623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.757646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.757660 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.762212 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.777260 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.791684 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.810271 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.824349 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.841738 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.860926 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.860996 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.861007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.861032 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.861045 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.863274 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:01Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.964364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.964409 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.964426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.964446 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:01 crc kubenswrapper[4725]: I1202 13:05:01.964475 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:01Z","lastTransitionTime":"2025-12-02T13:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.067270 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.067316 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.067327 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.067344 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.067355 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.100675 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.118287 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.136161 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.152570 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.166579 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.170390 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.170428 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.170438 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.170453 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.170480 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.185896 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.199530 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.211101 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.223131 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.239002 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.254601 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.265534 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.267626 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.267789 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.273403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.273435 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.273445 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.273473 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.273483 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.280524 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.290426 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.302449 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.315417 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.331943 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.376444 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.376507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.376517 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.376534 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.376544 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.412217 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vs5fz"] Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.412728 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.412818 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.425759 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.444687 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.447856 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.447896 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.447904 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.447923 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.447934 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.457717 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.459539 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.462812 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.462840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.462848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.462862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.462872 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.468843 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.473151 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.476479 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.476517 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.476529 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.476546 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.476557 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.480479 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.486914 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.489337 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.490646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.490685 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.490698 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.490716 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.490727 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.501423 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.503053 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.506480 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.506526 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.506557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.506577 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.506589 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.512821 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.512858 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbhzm\" (UniqueName: \"kubernetes.io/projected/9feadd7b-f199-41c3-b908-43774e1cd2f3-kube-api-access-wbhzm\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.516274 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.521581 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.521722 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.523261 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.523294 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.523305 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.523320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.523331 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.528692 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.538358 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.549520 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.557203 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" event={"ID":"e61071d0-15cb-43d9-a3f1-df802d821e1f","Type":"ContainerStarted","Data":"6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.557263 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" event={"ID":"e61071d0-15cb-43d9-a3f1-df802d821e1f","Type":"ContainerStarted","Data":"5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.557276 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" event={"ID":"e61071d0-15cb-43d9-a3f1-df802d821e1f","Type":"ContainerStarted","Data":"9b917a6767af7ae8fd29e17451776d3a321a8d042caa9a02cafbbe834594f29e"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.567307 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.584084 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.594244 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.604706 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.613177 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.613500 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbhzm\" (UniqueName: \"kubernetes.io/projected/9feadd7b-f199-41c3-b908-43774e1cd2f3-kube-api-access-wbhzm\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.613537 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.613906 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:02 crc kubenswrapper[4725]: E1202 13:05:02.614021 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:03.113971018 +0000 UTC m=+34.070612713 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.623048 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.626594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.626627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.626639 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.626657 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.626667 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.629354 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbhzm\" (UniqueName: \"kubernetes.io/projected/9feadd7b-f199-41c3-b908-43774e1cd2f3-kube-api-access-wbhzm\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.639954 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.649198 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.658788 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.671228 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.681546 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.690825 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.700290 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.708774 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.719419 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.728489 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.728531 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.728542 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.728558 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.728569 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.737633 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.749071 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.765452 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.778045 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.790938 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.801149 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.813003 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.823818 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:02Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.830801 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.830832 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.830843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.830856 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.830866 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.933306 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.933345 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.933356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.933372 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:02 crc kubenswrapper[4725]: I1202 13:05:02.933383 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:02Z","lastTransitionTime":"2025-12-02T13:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.035391 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.035452 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.035486 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.035516 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.035531 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.117872 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:03 crc kubenswrapper[4725]: E1202 13:05:03.117994 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:03 crc kubenswrapper[4725]: E1202 13:05:03.118056 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:04.118040086 +0000 UTC m=+35.074681781 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.137351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.137391 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.137401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.137418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.137431 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.239686 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.239730 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.239743 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.239760 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.239772 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.267215 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:03 crc kubenswrapper[4725]: E1202 13:05:03.267413 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.267576 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:03 crc kubenswrapper[4725]: E1202 13:05:03.267777 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.346156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.346244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.346290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.346317 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.346334 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.448872 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.448919 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.448936 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.448958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.448973 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.551535 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.551583 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.551596 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.551616 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.551630 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.653910 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.654173 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.654243 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.654326 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.654433 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.758200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.758271 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.758310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.758350 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.758416 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.861041 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.861078 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.861087 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.861102 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.861112 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.964597 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.964654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.964663 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.964681 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:03 crc kubenswrapper[4725]: I1202 13:05:03.964690 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:03Z","lastTransitionTime":"2025-12-02T13:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.068812 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.068905 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.068935 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.068989 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.069015 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.129911 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.130204 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:05:20.13015903 +0000 UTC m=+51.086800765 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.130743 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.130841 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.130873 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.130908 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.130937 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.130987 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131011 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131080 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:20.131062442 +0000 UTC m=+51.087704177 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131107 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131131 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131148 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131220 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:20.131198016 +0000 UTC m=+51.087839931 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131325 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131369 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:20.131348529 +0000 UTC m=+51.087990464 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.130941 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131525 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131557 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131652 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:20.131624146 +0000 UTC m=+51.088265861 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.131862 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:06.131841381 +0000 UTC m=+37.088483276 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.171799 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.171860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.171872 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.171892 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.171903 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.268159 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.268606 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.268167 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:04 crc kubenswrapper[4725]: E1202 13:05:04.268941 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.274015 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.274049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.274060 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.274074 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.274085 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.376605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.376646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.376654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.376670 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.376680 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.480178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.480217 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.480226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.480244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.480253 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.583157 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.583199 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.583209 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.583225 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.583236 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.685434 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.685486 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.685494 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.685511 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.685520 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.788038 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.788080 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.788092 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.788110 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.788120 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.890924 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.891183 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.891252 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.891325 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.891425 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.993634 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.994171 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.994234 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.994307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:04 crc kubenswrapper[4725]: I1202 13:05:04.994374 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:04Z","lastTransitionTime":"2025-12-02T13:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.096813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.097122 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.097218 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.097298 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.097391 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.199784 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.200414 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.200525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.200625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.200690 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.267885 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.268020 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:05 crc kubenswrapper[4725]: E1202 13:05:05.268275 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:05 crc kubenswrapper[4725]: E1202 13:05:05.268407 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.303615 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.303668 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.303683 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.303702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.303716 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.406524 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.406616 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.406629 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.406647 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.406659 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.509153 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.509196 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.509205 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.509225 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.509236 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.611572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.611608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.611618 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.611634 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.611644 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.714916 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.714955 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.714967 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.715007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.715035 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.817614 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.817658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.817675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.817700 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.817718 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.920474 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.920564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.920577 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.920594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:05 crc kubenswrapper[4725]: I1202 13:05:05.920603 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:05Z","lastTransitionTime":"2025-12-02T13:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.022928 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.022962 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.023009 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.023025 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.023035 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.125589 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.125644 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.125677 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.125713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.125731 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.152229 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:06 crc kubenswrapper[4725]: E1202 13:05:06.152386 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:06 crc kubenswrapper[4725]: E1202 13:05:06.152445 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:10.15242753 +0000 UTC m=+41.109069225 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.228129 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.228195 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.228216 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.228240 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.228253 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.267704 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.267172 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:06 crc kubenswrapper[4725]: E1202 13:05:06.267877 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:06 crc kubenswrapper[4725]: E1202 13:05:06.268051 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.331863 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.331906 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.331914 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.331929 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.331939 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.434316 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.434369 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.434382 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.434401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.434415 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.536636 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.536679 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.536690 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.536706 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.536718 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.638913 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.638965 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.638976 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.638990 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.639000 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.742045 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.742107 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.742132 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.742160 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.742184 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.845436 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.845551 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.845584 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.845616 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.845637 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.948185 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.948230 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.948239 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.948255 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:06 crc kubenswrapper[4725]: I1202 13:05:06.948266 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:06Z","lastTransitionTime":"2025-12-02T13:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.050725 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.050768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.050784 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.050805 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.050817 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.153063 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.153097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.153112 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.153130 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.153142 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.256207 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.256244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.256256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.256274 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.256285 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.267695 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.267706 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:07 crc kubenswrapper[4725]: E1202 13:05:07.267826 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:07 crc kubenswrapper[4725]: E1202 13:05:07.267877 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.358338 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.358389 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.358402 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.358419 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.358439 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.460999 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.461048 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.461058 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.461075 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.461086 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.564112 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.564154 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.564162 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.564178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.564187 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.666511 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.666565 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.666576 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.666594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.666606 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.768654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.768713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.768722 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.768736 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.768746 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.870944 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.870977 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.871028 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.871046 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.871088 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.973415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.973555 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.973579 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.973607 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:07 crc kubenswrapper[4725]: I1202 13:05:07.973658 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:07Z","lastTransitionTime":"2025-12-02T13:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.076568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.076612 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.076623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.076641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.076653 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.179076 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.179130 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.179143 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.179164 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.179175 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.268066 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.268208 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:08 crc kubenswrapper[4725]: E1202 13:05:08.268270 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:08 crc kubenswrapper[4725]: E1202 13:05:08.268447 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.281743 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.281800 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.281818 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.281845 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.281864 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.387517 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.387740 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.387934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.387949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.387959 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.489973 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.490015 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.490027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.490045 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.490056 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.592539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.592579 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.592592 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.592609 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.592620 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.695433 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.695526 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.695549 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.695577 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.695600 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.797416 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.797494 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.797507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.797528 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.797541 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.900304 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.900345 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.900354 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.900370 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:08 crc kubenswrapper[4725]: I1202 13:05:08.900379 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:08Z","lastTransitionTime":"2025-12-02T13:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.003132 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.003187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.003200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.003222 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.003234 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.105844 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.105887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.105899 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.105920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.105935 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.208658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.209083 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.209247 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.209360 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.209451 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.267751 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:09 crc kubenswrapper[4725]: E1202 13:05:09.267937 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.268277 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:09 crc kubenswrapper[4725]: E1202 13:05:09.269008 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.287167 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.304693 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.312717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.312770 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.312783 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.312802 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.312815 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.326044 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.336309 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.352021 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.361962 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.379307 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.404757 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.415797 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.415863 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.415924 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.415960 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.415978 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.434457 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.450070 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.464341 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.482117 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.497698 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.510690 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.517737 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.517772 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.517782 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.517798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.517808 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.527577 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.543631 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.554341 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.620141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.620175 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.620186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.620209 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.620228 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.722741 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.722778 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.722788 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.722803 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.722812 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.825397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.825723 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.825837 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.825970 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.826141 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.928265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.928310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.928321 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.928338 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:09 crc kubenswrapper[4725]: I1202 13:05:09.928350 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:09Z","lastTransitionTime":"2025-12-02T13:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.030820 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.030864 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.030873 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.030953 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.030965 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.133480 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.133517 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.133530 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.133549 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.133560 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.193639 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:10 crc kubenswrapper[4725]: E1202 13:05:10.193826 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:10 crc kubenswrapper[4725]: E1202 13:05:10.193951 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:18.193928274 +0000 UTC m=+49.150570019 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.235786 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.235832 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.235848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.235870 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.235896 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.267450 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.267518 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:10 crc kubenswrapper[4725]: E1202 13:05:10.267655 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:10 crc kubenswrapper[4725]: E1202 13:05:10.267945 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.338969 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.339794 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.339810 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.339825 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.339833 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.442957 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.443003 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.443013 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.443029 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.443040 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.545274 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.545563 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.545690 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.545796 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.545877 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.647994 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.648027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.648039 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.648056 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.648069 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.750703 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.750740 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.750749 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.750766 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.750776 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.854282 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.854359 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.854385 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.854508 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.854543 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.957227 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.957274 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.957284 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.957301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:10 crc kubenswrapper[4725]: I1202 13:05:10.957313 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:10Z","lastTransitionTime":"2025-12-02T13:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.060478 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.060526 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.060537 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.060555 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.060567 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.163440 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.163497 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.163506 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.163528 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.163553 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.266127 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.266168 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.266178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.266193 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.266203 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.267486 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.267532 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:11 crc kubenswrapper[4725]: E1202 13:05:11.267577 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:11 crc kubenswrapper[4725]: E1202 13:05:11.267716 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.368013 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.368075 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.368092 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.368121 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.368139 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.474885 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.474944 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.474961 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.474984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.475001 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.577524 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.577582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.577599 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.577625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.577643 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.679671 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.679702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.679710 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.679723 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.679732 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.782251 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.782302 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.782319 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.782339 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.782348 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.884502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.884580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.884617 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.884637 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.884649 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.986871 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.986920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.986929 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.986942 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:11 crc kubenswrapper[4725]: I1202 13:05:11.986951 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:11Z","lastTransitionTime":"2025-12-02T13:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.089415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.089495 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.089510 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.089531 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.089543 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.191369 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.191414 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.191427 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.191447 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.191480 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.267810 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.267885 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.267966 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.268074 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.294050 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.294099 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.294109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.294125 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.294134 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.396839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.396883 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.396897 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.396913 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.396925 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.499117 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.499150 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.499163 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.499180 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.499189 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.601353 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.601401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.601412 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.601429 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.601443 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.703526 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.703578 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.703589 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.703606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.703617 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.806651 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.806705 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.806716 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.806734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.806745 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.836339 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.836386 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.836398 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.836419 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.836437 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.849350 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:12Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.853211 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.853260 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.853269 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.853284 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.853295 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.864386 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:12Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.867841 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.867893 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.867903 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.867920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.867948 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.877412 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:12Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.881022 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.881058 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.881068 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.881084 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.881093 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.894183 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:12Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.897743 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.897796 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.897814 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.897837 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.897855 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.908140 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:12Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:12 crc kubenswrapper[4725]: E1202 13:05:12.908258 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.909651 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.909678 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.909687 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.909706 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:12 crc kubenswrapper[4725]: I1202 13:05:12.909717 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:12Z","lastTransitionTime":"2025-12-02T13:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.011426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.011533 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.011553 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.011582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.011603 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.114074 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.114130 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.114145 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.114164 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.114180 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.217807 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.217860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.217869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.217884 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.217894 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.268283 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.268305 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:13 crc kubenswrapper[4725]: E1202 13:05:13.268743 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:13 crc kubenswrapper[4725]: E1202 13:05:13.269036 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.324238 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.324296 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.324305 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.324319 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.324328 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.427148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.428284 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.428531 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.428734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.428920 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.531173 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.531520 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.531611 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.531729 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.531813 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.635643 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.635963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.636042 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.636124 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.636203 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.738629 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.738904 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.738994 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.739162 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.739266 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.842311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.842736 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.842827 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.842934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.843027 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.945848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.945905 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.945921 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.945945 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:13 crc kubenswrapper[4725]: I1202 13:05:13.945962 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:13Z","lastTransitionTime":"2025-12-02T13:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.048777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.048817 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.048828 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.048842 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.048852 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.151668 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.151705 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.151714 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.151729 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.151738 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.253817 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.253885 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.253894 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.253915 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.253928 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.267163 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:14 crc kubenswrapper[4725]: E1202 13:05:14.267271 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.267163 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:14 crc kubenswrapper[4725]: E1202 13:05:14.267527 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.356764 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.356811 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.356824 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.356842 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.356856 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.458888 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.458927 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.458937 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.458954 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.458964 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.561288 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.561338 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.561351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.561371 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.561386 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.663773 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.663806 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.663814 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.663827 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.663836 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.766284 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.766323 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.766337 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.766356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.766371 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.868717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.868789 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.868808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.868836 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.868853 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.971358 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.971401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.971411 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.971431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:14 crc kubenswrapper[4725]: I1202 13:05:14.971442 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:14Z","lastTransitionTime":"2025-12-02T13:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.074000 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.074054 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.074067 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.074090 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.074103 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.176914 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.176953 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.176965 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.176983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.176995 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.268032 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:15 crc kubenswrapper[4725]: E1202 13:05:15.268251 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.268349 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:15 crc kubenswrapper[4725]: E1202 13:05:15.268776 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.269081 4725 scope.go:117] "RemoveContainer" containerID="73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.278949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.278979 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.278989 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.279003 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.279013 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.381273 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.381308 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.381319 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.381333 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.381343 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.494768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.495658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.495708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.495732 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.495746 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.606065 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.606105 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.606117 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.606132 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.606146 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.608559 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/1.log" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.611290 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.611447 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.625267 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.638779 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.654234 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.672927 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.685829 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.696722 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.708095 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.709140 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.709339 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.709479 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.709634 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.709749 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.722567 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.741680 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.753359 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.764725 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.774726 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.786953 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.798082 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.812298 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.813097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.813131 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.813141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.813157 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.813166 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.829401 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.843724 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:15Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.915940 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.916532 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.916647 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.916763 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:15 crc kubenswrapper[4725]: I1202 13:05:15.916863 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:15Z","lastTransitionTime":"2025-12-02T13:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.019441 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.019798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.019999 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.020093 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.020193 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.122054 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.122097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.122106 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.122119 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.122130 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.224386 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.224434 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.224446 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.224489 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.224501 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.267744 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:16 crc kubenswrapper[4725]: E1202 13:05:16.268190 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.267744 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:16 crc kubenswrapper[4725]: E1202 13:05:16.268490 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.329248 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.329325 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.329365 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.329388 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.329403 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.433070 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.433128 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.433138 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.433156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.433167 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.536238 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.536315 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.536340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.536380 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.536408 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.618116 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/2.log" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.619177 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/1.log" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.622169 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b" exitCode=1 Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.622265 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.622319 4725 scope.go:117] "RemoveContainer" containerID="73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.623166 4725 scope.go:117] "RemoveContainer" containerID="efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b" Dec 02 13:05:16 crc kubenswrapper[4725]: E1202 13:05:16.623382 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.639407 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.639672 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.639802 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.639912 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.640014 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.642828 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.655907 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.685550 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.698586 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.715653 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.732507 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.742236 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.742302 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.742316 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.742340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.742356 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.746017 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.767643 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.800057 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.820320 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.834585 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.845438 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.845537 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.845565 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.845589 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.845608 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.854362 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.872171 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.888642 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.904961 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.923081 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.934361 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:16Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.948555 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.948606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.948617 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.948636 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:16 crc kubenswrapper[4725]: I1202 13:05:16.948649 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:16Z","lastTransitionTime":"2025-12-02T13:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.051442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.051545 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.051587 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.051631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.051659 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.154582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.154657 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.154675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.154701 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.154720 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.257491 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.257562 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.257579 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.257605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.257624 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.268087 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:17 crc kubenswrapper[4725]: E1202 13:05:17.268276 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.268718 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:17 crc kubenswrapper[4725]: E1202 13:05:17.268925 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.360089 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.360139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.360149 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.360166 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.360184 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.462648 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.462699 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.462710 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.462730 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.462745 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.565848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.565885 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.565896 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.565912 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.565923 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.628072 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/2.log" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.669010 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.669040 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.669049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.669062 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.669070 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.772692 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.772768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.772787 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.772812 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.772828 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.875314 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.875364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.875377 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.875397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.875411 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.977934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.977973 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.977985 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.978003 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:17 crc kubenswrapper[4725]: I1202 13:05:17.978016 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:17Z","lastTransitionTime":"2025-12-02T13:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.047071 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.059784 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.073098 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.082502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.082574 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.082587 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.082605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.082641 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.096060 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73ea125130132b14bd027d9c092b03f6c94b6c4dbc82d4d17227049f50036cea\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:04:59Z\\\",\\\"message\\\":\\\"webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:04:59Z is after 2025-08-24T17:21:41Z]\\\\nI1202 13:04:59.371353 6157 services_controller.go:434] Service openshift-operator-lifecycle-manager/olm-operator-metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{olm-operator-metrics openshift-operator-lifecycle-manager 6ee2ddfb-7cb2-4c1f-895d-c8afbc5f6db6 4666 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:olm-operator] map[capability.openshift.io/name:OperatorLifecycleManager include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:olm-operator-serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0073771ff \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https-metrics,Protocol:TCP,Port:8443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: olm-operator,},ClusterIP:10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.107560 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.120977 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.131905 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.145864 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.157913 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.169759 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.180381 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.184487 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.184512 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.184520 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.184537 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.184546 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.195225 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.214322 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.225139 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.235295 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.244536 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.254685 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.263516 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.267654 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.267654 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:18 crc kubenswrapper[4725]: E1202 13:05:18.267768 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:18 crc kubenswrapper[4725]: E1202 13:05:18.267865 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.275387 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.282857 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:18 crc kubenswrapper[4725]: E1202 13:05:18.283021 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:18 crc kubenswrapper[4725]: E1202 13:05:18.283081 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:34.283064461 +0000 UTC m=+65.239706156 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.287311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.287359 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.287370 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.287386 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.287396 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.389757 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.389796 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.389806 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.389823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.389834 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.406523 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.408033 4725 scope.go:117] "RemoveContainer" containerID="efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b" Dec 02 13:05:18 crc kubenswrapper[4725]: E1202 13:05:18.408337 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.424947 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.442799 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.457162 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.466259 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.478395 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.488243 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.491587 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.491642 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.491659 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.491680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.491695 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.501192 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.512036 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.524924 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.544453 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.554245 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.564596 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.575259 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.587451 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.594207 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.594250 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.594262 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.594279 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.594292 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.607439 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.620641 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.631871 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.645594 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:18Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.696372 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.696433 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.696452 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.696610 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.696623 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.799240 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.799363 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.799418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.799456 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.799525 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.902852 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.903243 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.903361 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.903484 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:18 crc kubenswrapper[4725]: I1202 13:05:18.903607 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:18Z","lastTransitionTime":"2025-12-02T13:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.006097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.006352 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.006476 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.006557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.006621 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.108813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.109075 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.109158 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.109254 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.109338 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.212277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.212318 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.212328 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.212343 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.212353 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.267927 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:19 crc kubenswrapper[4725]: E1202 13:05:19.268050 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.268405 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:19 crc kubenswrapper[4725]: E1202 13:05:19.268538 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.281109 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.294749 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.306783 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.314561 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.314599 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.314609 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.314625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.314634 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.317110 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.329180 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.338892 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.351757 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.363505 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.381359 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.400148 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.410254 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.416226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.416254 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.416264 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.416277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.416287 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.419792 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.428249 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.439303 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.456632 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.467479 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.481003 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.493139 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.517641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.517688 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.517702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.517717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.517728 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.619858 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.619897 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.619907 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.619923 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.619933 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.722508 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.722547 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.722557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.722572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.722582 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.825294 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.825335 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.825344 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.825361 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.825370 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.927320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.927364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.927374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.927389 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:19 crc kubenswrapper[4725]: I1202 13:05:19.927399 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:19Z","lastTransitionTime":"2025-12-02T13:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.030060 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.030112 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.030124 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.030141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.030153 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.131956 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.131994 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.132005 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.132020 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.132029 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.205071 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205179 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:05:52.205163761 +0000 UTC m=+83.161805446 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.205283 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.205306 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205412 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205425 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205435 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205504 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:52.205495789 +0000 UTC m=+83.162137484 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205633 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205673 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205686 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205742 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:52.205727225 +0000 UTC m=+83.162368920 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.205872 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205966 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.205999 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:52.205992531 +0000 UTC m=+83.162634226 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.206024 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.206069 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.206093 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:05:52.206087384 +0000 UTC m=+83.162729079 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.234708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.234777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.234789 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.234808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.234820 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.267139 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.267168 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.267281 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:20 crc kubenswrapper[4725]: E1202 13:05:20.267368 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.337300 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.337340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.337351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.337364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.337373 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.440152 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.440378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.440390 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.440404 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.440413 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.542941 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.543189 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.543287 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.543368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.543436 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.646606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.646665 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.646682 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.646703 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.646719 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.749511 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.749548 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.749560 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.749576 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.749586 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.852182 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.852234 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.852244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.852259 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.852270 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.954324 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.954356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.954364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.954376 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:20 crc kubenswrapper[4725]: I1202 13:05:20.954385 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:20Z","lastTransitionTime":"2025-12-02T13:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.057572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.057622 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.057640 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.057663 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.057680 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.159643 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.159682 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.159691 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.159707 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.159719 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.262127 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.262165 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.262173 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.262185 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.262196 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.268339 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:21 crc kubenswrapper[4725]: E1202 13:05:21.268481 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.268810 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:21 crc kubenswrapper[4725]: E1202 13:05:21.268978 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.364690 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.364726 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.364735 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.364751 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.364761 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.467805 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.467846 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.467855 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.467870 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.467879 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.571073 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.571158 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.571167 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.571182 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.571191 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.673931 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.673998 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.674017 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.674042 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.674060 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.776820 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.776856 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.776865 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.776878 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.776887 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.879580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.879627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.879640 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.879659 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.879669 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.982049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.982090 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.982098 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.982113 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:21 crc kubenswrapper[4725]: I1202 13:05:21.982127 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:21Z","lastTransitionTime":"2025-12-02T13:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.085249 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.085317 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.085329 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.085348 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.085365 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.188006 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.188044 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.188053 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.188067 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.188077 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.267309 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.267336 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:22 crc kubenswrapper[4725]: E1202 13:05:22.267439 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:22 crc kubenswrapper[4725]: E1202 13:05:22.267585 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.290440 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.290504 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.290514 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.290528 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.290538 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.392813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.392853 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.392862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.392880 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.392890 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.496682 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.496744 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.496755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.496772 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.496785 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.599498 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.599569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.599580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.599606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.599621 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.702369 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.702416 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.702429 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.702444 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.702484 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.805814 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.805896 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.805920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.805959 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.805983 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.909125 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.909174 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.909186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.909203 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.909216 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.937560 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.937599 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.937608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.937622 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.937631 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: E1202 13:05:22.952054 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:22Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.955588 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.955653 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.955664 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.955686 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.955699 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: E1202 13:05:22.970214 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:22Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.975108 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.975182 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.975204 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.975232 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:22 crc kubenswrapper[4725]: I1202 13:05:22.975248 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:22Z","lastTransitionTime":"2025-12-02T13:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:22 crc kubenswrapper[4725]: E1202 13:05:22.999355 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:22Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.003911 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.004011 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.004040 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.004070 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.004091 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: E1202 13:05:23.023394 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:23Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.027934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.027968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.027978 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.027996 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.028012 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: E1202 13:05:23.042401 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:23Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:23 crc kubenswrapper[4725]: E1202 13:05:23.042548 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.044119 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.044150 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.044164 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.044180 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.044192 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.146561 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.146605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.146616 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.146630 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.146639 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.249188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.249226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.249235 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.249252 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.249262 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.267939 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.267986 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:23 crc kubenswrapper[4725]: E1202 13:05:23.268090 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:23 crc kubenswrapper[4725]: E1202 13:05:23.268153 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.351143 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.351195 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.351206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.351221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.351230 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.453670 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.453708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.453717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.453731 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.453740 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.556523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.556792 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.556854 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.556957 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.557020 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.658847 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.658887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.658897 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.658918 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.658929 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.762603 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.762645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.762655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.762669 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.762680 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.866567 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.866609 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.866620 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.866635 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.866643 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.970218 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.970261 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.970273 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.970294 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:23 crc kubenswrapper[4725]: I1202 13:05:23.970314 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:23Z","lastTransitionTime":"2025-12-02T13:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.073064 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.073096 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.073105 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.073118 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.073126 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.175840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.176644 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.176676 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.176708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.176731 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.267185 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.267226 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:24 crc kubenswrapper[4725]: E1202 13:05:24.267322 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:24 crc kubenswrapper[4725]: E1202 13:05:24.267431 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.279056 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.279094 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.279107 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.279125 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.279138 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.381206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.381243 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.381253 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.381270 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.381281 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.483772 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.483812 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.483823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.483839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.483892 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.588112 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.588165 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.588177 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.588194 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.588207 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.690289 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.690326 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.690334 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.690349 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.690358 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.792590 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.792652 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.792669 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.792692 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.792702 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.895400 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.895507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.895535 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.895566 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.895589 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.998541 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.998594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.998608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.998627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:24 crc kubenswrapper[4725]: I1202 13:05:24.998641 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:24Z","lastTransitionTime":"2025-12-02T13:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.101603 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.101680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.101704 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.101740 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.101762 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.205322 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.205393 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.205412 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.205438 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.205494 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.267544 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:25 crc kubenswrapper[4725]: E1202 13:05:25.267673 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.267759 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:25 crc kubenswrapper[4725]: E1202 13:05:25.268000 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.307877 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.307923 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.307935 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.307954 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.307966 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.410370 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.410402 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.410412 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.410426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.410437 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.513276 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.513325 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.513339 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.513359 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.513372 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.615606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.615646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.615658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.615675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.615689 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.718477 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.718521 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.718529 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.718542 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.718551 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.820932 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.820976 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.820984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.821000 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.821009 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.923208 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.923244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.923254 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.923276 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:25 crc kubenswrapper[4725]: I1202 13:05:25.923297 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:25Z","lastTransitionTime":"2025-12-02T13:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.025935 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.025979 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.026010 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.026026 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.026038 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.128214 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.128254 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.128274 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.128311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.128322 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.232525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.232574 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.232594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.232612 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.232625 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.267334 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:26 crc kubenswrapper[4725]: E1202 13:05:26.267439 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.267341 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:26 crc kubenswrapper[4725]: E1202 13:05:26.267574 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.334898 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.334928 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.334937 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.334951 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.334961 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.437593 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.437631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.437641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.437657 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.437666 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.541034 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.541119 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.541133 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.541186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.541199 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.643557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.643621 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.643633 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.643646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.643655 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.745883 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.745911 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.745919 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.745933 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.745942 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.848436 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.848523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.848533 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.848546 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.848555 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.950949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.950983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.950990 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.951004 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:26 crc kubenswrapper[4725]: I1202 13:05:26.951012 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:26Z","lastTransitionTime":"2025-12-02T13:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.053696 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.053962 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.054037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.054123 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.054186 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.156797 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.156833 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.156844 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.156860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.156872 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.259016 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.259042 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.259050 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.259063 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.259071 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.267189 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.267233 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:27 crc kubenswrapper[4725]: E1202 13:05:27.267547 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:27 crc kubenswrapper[4725]: E1202 13:05:27.267696 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.361121 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.361155 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.361166 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.361181 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.361211 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.463538 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.463763 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.463808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.464136 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.464155 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.566681 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.566945 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.567164 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.567295 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.567404 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.669651 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.669680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.669690 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.669702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.669711 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.772005 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.772049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.772058 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.772073 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.772081 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.874685 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.874727 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.874735 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.874749 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.874758 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.976699 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.976753 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.976764 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.976811 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:27 crc kubenswrapper[4725]: I1202 13:05:27.976824 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:27Z","lastTransitionTime":"2025-12-02T13:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.079240 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.079545 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.079568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.079586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.079596 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.182143 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.182192 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.182201 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.182215 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.182224 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.267963 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.268006 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:28 crc kubenswrapper[4725]: E1202 13:05:28.268097 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:28 crc kubenswrapper[4725]: E1202 13:05:28.268179 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.284775 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.284850 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.284915 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.284945 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.284967 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.387244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.387297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.387312 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.387329 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.387338 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.489446 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.489520 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.489530 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.489544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.489553 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.592216 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.592289 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.592305 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.592329 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.592342 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.695194 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.695223 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.695231 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.695244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.695253 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.797502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.797572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.797592 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.797621 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.797641 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.900229 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.900297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.900307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.900321 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:28 crc kubenswrapper[4725]: I1202 13:05:28.900330 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:28Z","lastTransitionTime":"2025-12-02T13:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.006182 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.006260 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.006276 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.006314 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.006332 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.109139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.109177 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.109188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.109206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.109216 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.211396 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.211451 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.211483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.211524 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.211536 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.267195 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:29 crc kubenswrapper[4725]: E1202 13:05:29.267324 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.267554 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:29 crc kubenswrapper[4725]: E1202 13:05:29.267605 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.279683 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.292700 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.302865 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.314529 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.314608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.314637 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.314646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.314661 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.314670 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.324746 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.335399 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.345675 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.358219 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.375805 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.386214 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.396916 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.406309 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.416627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.416662 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.416671 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.416685 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.416697 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.416621 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.427170 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.443032 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.454195 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.468235 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.487864 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:29Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.518550 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.518608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.518627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.518641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.518650 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.621540 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.621877 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.621888 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.621904 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.621916 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.724368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.724409 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.724420 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.724437 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.724476 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.826719 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.826751 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.826761 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.826774 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.826783 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.928893 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.928933 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.928943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.928958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:29 crc kubenswrapper[4725]: I1202 13:05:29.928970 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:29Z","lastTransitionTime":"2025-12-02T13:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.031112 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.031418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.031656 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.031777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.031855 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.136098 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.136137 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.136146 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.136160 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.136169 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.238408 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.238440 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.238449 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.238490 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.238500 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.267145 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.267213 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:30 crc kubenswrapper[4725]: E1202 13:05:30.267319 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:30 crc kubenswrapper[4725]: E1202 13:05:30.267491 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.340824 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.340856 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.340864 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.340876 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.340886 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.443639 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.443669 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.443678 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.443691 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.443700 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.545788 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.545839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.545848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.545861 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.545869 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.647723 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.647758 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.647766 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.647780 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.647789 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.750089 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.750156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.750168 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.750182 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.750192 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.852633 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.852666 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.852680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.852695 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.852704 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.954798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.954855 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.954866 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.954880 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:30 crc kubenswrapper[4725]: I1202 13:05:30.954890 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:30Z","lastTransitionTime":"2025-12-02T13:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.057333 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.057368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.057379 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.057395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.057405 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.160223 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.160272 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.160290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.160306 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.160316 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.262389 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.262417 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.262426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.262439 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.262448 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.267765 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:31 crc kubenswrapper[4725]: E1202 13:05:31.267950 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.268367 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:31 crc kubenswrapper[4725]: E1202 13:05:31.268496 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.365186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.365243 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.365256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.365275 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.365287 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.477509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.477549 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.477559 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.477575 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.477586 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.580362 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.580401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.580412 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.580428 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.580440 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.682916 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.683000 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.683022 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.683056 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.683079 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.785723 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.785841 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.785854 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.785869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.785877 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.887978 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.888014 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.888024 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.888037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.888047 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.989962 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.989999 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.990009 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.990024 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:31 crc kubenswrapper[4725]: I1202 13:05:31.990033 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:31Z","lastTransitionTime":"2025-12-02T13:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.091766 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.091805 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.091825 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.091842 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.091854 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.194120 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.194180 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.194192 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.194207 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.194218 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.267996 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.268079 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:32 crc kubenswrapper[4725]: E1202 13:05:32.268128 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:32 crc kubenswrapper[4725]: E1202 13:05:32.268200 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.268779 4725 scope.go:117] "RemoveContainer" containerID="efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b" Dec 02 13:05:32 crc kubenswrapper[4725]: E1202 13:05:32.268994 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.295975 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.296022 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.296033 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.296049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.296063 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.398394 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.398441 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.398451 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.398483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.398495 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.500940 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.501000 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.501010 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.501028 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.501037 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.603351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.603387 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.603395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.603409 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.603418 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.705683 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.705730 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.705747 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.705800 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.705832 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.808394 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.808424 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.808432 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.808446 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.808481 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.911037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.911077 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.911088 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.911109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:32 crc kubenswrapper[4725]: I1202 13:05:32.911120 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:32Z","lastTransitionTime":"2025-12-02T13:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.013960 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.014025 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.014038 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.014063 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.014078 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.117157 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.117221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.117234 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.117257 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.117272 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.219701 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.219744 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.219755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.219770 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.219781 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.267331 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.267398 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.267484 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.267581 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.271742 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.271767 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.271775 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.271789 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.271799 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.283072 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:33Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.286841 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.286865 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.286874 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.286887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.286896 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.298193 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:33Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.301234 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.301264 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.301273 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.301288 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.301297 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.312356 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:33Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.315282 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.315312 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.315324 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.315340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.315351 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.327264 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:33Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.332703 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.332761 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.332771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.332786 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.332795 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.345236 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:33Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:33 crc kubenswrapper[4725]: E1202 13:05:33.345350 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.346782 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.346831 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.346840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.346853 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.346863 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.448683 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.448715 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.448724 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.448738 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.448769 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.550684 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.550712 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.550720 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.550734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.550742 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.652767 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.652800 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.652808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.652822 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.652831 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.756239 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.756612 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.756761 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.756875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.757053 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.859249 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.859281 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.859291 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.859307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.859318 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.961410 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.961452 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.961483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.961501 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:33 crc kubenswrapper[4725]: I1202 13:05:33.961512 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:33Z","lastTransitionTime":"2025-12-02T13:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.063819 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.063853 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.063866 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.063882 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.063894 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.166206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.166254 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.166265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.166283 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.166299 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.267384 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.267384 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:34 crc kubenswrapper[4725]: E1202 13:05:34.267612 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:34 crc kubenswrapper[4725]: E1202 13:05:34.267818 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.269862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.269916 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.269959 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.269983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.270003 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.328755 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:34 crc kubenswrapper[4725]: E1202 13:05:34.328947 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:34 crc kubenswrapper[4725]: E1202 13:05:34.329041 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:06:06.329016543 +0000 UTC m=+97.285658238 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.372712 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.372787 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.372809 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.372839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.372859 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.475579 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.475632 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.475645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.475667 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.475679 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.578260 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.578320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.578338 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.578364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.578384 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.680976 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.681008 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.681016 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.681031 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.681039 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.783869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.783900 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.783912 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.783928 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.783939 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.886594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.886631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.886640 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.886656 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.886668 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.990096 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.990145 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.990154 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.990172 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:34 crc kubenswrapper[4725]: I1202 13:05:34.990183 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:34Z","lastTransitionTime":"2025-12-02T13:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.092724 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.094036 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.094144 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.094249 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.094335 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.196952 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.197200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.197267 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.197333 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.197389 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.269817 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.270137 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:35 crc kubenswrapper[4725]: E1202 13:05:35.270260 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:35 crc kubenswrapper[4725]: E1202 13:05:35.270396 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.299958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.300019 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.300041 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.300058 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.300073 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.402188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.402226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.402240 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.402255 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.402265 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.505877 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.506104 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.506200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.506264 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.506329 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.608715 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.608742 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.608750 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.608764 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.608805 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.687720 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/0.log" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.687986 4725 generic.go:334] "Generic (PLEG): container finished" podID="4a8c02a6-36ad-4a9c-88b0-064dcc9b7327" containerID="7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463" exitCode=1 Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.688119 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerDied","Data":"7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.688656 4725 scope.go:117] "RemoveContainer" containerID="7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.702431 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.710873 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.711152 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.711160 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.711173 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.711183 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.716342 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.728968 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.748111 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.760409 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.770731 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.784158 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.797070 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.810550 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.816256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.816296 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.816307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.816327 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.816339 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.827016 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.838739 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.852961 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.863419 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.874823 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.886048 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.900238 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.919373 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.919403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.919413 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.919428 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.919439 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:35Z","lastTransitionTime":"2025-12-02T13:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.923489 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:35 crc kubenswrapper[4725]: I1202 13:05:35.933910 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:35Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.021446 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.021496 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.021507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.021521 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.021531 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.123790 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.123822 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.123834 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.123850 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.123861 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.226659 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.226703 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.226715 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.226732 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.226745 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.268587 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:36 crc kubenswrapper[4725]: E1202 13:05:36.268673 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.269357 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:36 crc kubenswrapper[4725]: E1202 13:05:36.269509 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.278592 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.328996 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.329025 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.329034 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.329049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.329058 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.431516 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.431569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.431582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.431597 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.431606 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.534768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.534815 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.534826 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.534842 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.534854 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.637472 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.637520 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.637533 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.637575 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.637587 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.692152 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/0.log" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.692660 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerStarted","Data":"9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.704434 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.716576 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.730826 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.740337 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.740575 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.740640 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.740713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.740776 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.750411 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.765890 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.777560 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.790622 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.804979 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.819105 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.834214 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.843275 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.843330 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.843339 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.843354 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.843367 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.849217 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.862549 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.874567 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.887839 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.900447 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.919992 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.937091 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.945224 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.945273 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.945283 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.945296 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.945307 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:36Z","lastTransitionTime":"2025-12-02T13:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.947857 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:36 crc kubenswrapper[4725]: I1202 13:05:36.958372 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:36Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.047097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.047139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.047148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.047163 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.047172 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.149218 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.149242 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.149250 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.149262 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.149270 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.251935 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.251962 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.251970 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.252019 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.252030 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.267788 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.267860 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:37 crc kubenswrapper[4725]: E1202 13:05:37.267888 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:37 crc kubenswrapper[4725]: E1202 13:05:37.267922 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.354018 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.354052 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.354061 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.354075 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.354085 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.456411 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.456448 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.456472 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.456488 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.456497 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.558659 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.558688 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.558696 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.558709 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.558718 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.660986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.661032 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.661042 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.661057 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.661066 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.763214 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.763244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.763253 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.763269 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.763279 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.865769 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.865833 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.865843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.865860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.865872 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.967642 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.967680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.967689 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.967713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:37 crc kubenswrapper[4725]: I1202 13:05:37.967724 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:37Z","lastTransitionTime":"2025-12-02T13:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.069912 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.069988 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.070005 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.070030 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.070049 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.172278 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.172349 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.172364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.172384 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.172396 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.267980 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.268027 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:38 crc kubenswrapper[4725]: E1202 13:05:38.268098 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:38 crc kubenswrapper[4725]: E1202 13:05:38.268167 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.274240 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.274269 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.274279 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.274295 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.274306 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.376327 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.376364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.376376 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.376394 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.376405 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.479697 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.479787 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.479817 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.479849 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.479871 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.582085 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.582271 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.582297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.582372 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.582394 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.684503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.684551 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.684563 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.684581 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.684594 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.786514 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.786553 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.786564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.786581 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.786592 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.889034 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.889063 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.889070 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.889085 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.889093 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.991683 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.991732 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.991746 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.991763 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:38 crc kubenswrapper[4725]: I1202 13:05:38.991775 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:38Z","lastTransitionTime":"2025-12-02T13:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.093688 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.093725 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.093737 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.093753 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.093764 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.196382 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.196445 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.196454 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.196483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.196492 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.267654 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:39 crc kubenswrapper[4725]: E1202 13:05:39.267773 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.267963 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:39 crc kubenswrapper[4725]: E1202 13:05:39.268117 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.285923 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.295811 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.298509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.298541 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.298552 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.298566 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.298574 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.308099 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.318935 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.331598 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.344003 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.353179 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.362327 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.373046 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.386577 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.400014 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.400072 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.400085 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.400103 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.400114 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.405279 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.414609 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.424687 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.437165 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.449396 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.465287 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.485799 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.497103 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.501483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.501508 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.501516 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.501529 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.501540 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.508129 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:39Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.603859 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.604126 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.604202 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.604277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.604338 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.706032 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.706065 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.706073 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.706087 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.706098 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.808340 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.808370 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.808378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.808391 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.808400 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.911056 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.911490 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.911768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.911917 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:39 crc kubenswrapper[4725]: I1202 13:05:39.912055 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:39Z","lastTransitionTime":"2025-12-02T13:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.014443 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.014682 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.014748 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.014840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.014898 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.118052 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.118347 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.118449 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.118538 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.118602 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.220722 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.221310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.221402 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.221535 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.221639 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.267724 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.267815 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:40 crc kubenswrapper[4725]: E1202 13:05:40.267977 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:40 crc kubenswrapper[4725]: E1202 13:05:40.270639 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.323410 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.323486 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.323503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.323525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.323540 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.425523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.425564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.425574 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.425589 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.425599 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.527910 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.528176 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.528244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.528303 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.528362 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.631189 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.631221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.631229 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.631242 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.631251 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.733593 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.733625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.733633 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.733646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.733656 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.835697 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.835741 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.835753 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.835775 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.835785 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.938594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.938641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.938654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.938674 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:40 crc kubenswrapper[4725]: I1202 13:05:40.938688 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:40Z","lastTransitionTime":"2025-12-02T13:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.040773 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.040816 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.040826 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.040861 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.040875 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.143010 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.143044 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.143053 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.143067 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.143078 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.245857 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.245891 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.245901 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.245915 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.245927 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.267229 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:41 crc kubenswrapper[4725]: E1202 13:05:41.267328 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.267241 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:41 crc kubenswrapper[4725]: E1202 13:05:41.267418 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.348365 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.348406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.348415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.348429 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.348490 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.451649 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.451693 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.451704 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.451722 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.451735 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.556706 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.556755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.556766 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.556786 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.556800 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.659569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.659631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.659643 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.659684 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.659698 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.761872 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.761910 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.761922 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.761939 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.761955 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.864535 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.864564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.864572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.864589 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.864605 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.967248 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.967298 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.967311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.967326 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:41 crc kubenswrapper[4725]: I1202 13:05:41.967334 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:41Z","lastTransitionTime":"2025-12-02T13:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.069072 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.069101 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.069110 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.069124 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.069132 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.171437 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.171496 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.171507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.171522 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.171531 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.267347 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.267376 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:42 crc kubenswrapper[4725]: E1202 13:05:42.267487 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:42 crc kubenswrapper[4725]: E1202 13:05:42.267658 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.273978 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.274014 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.274026 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.274040 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.274050 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.376301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.376455 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.376483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.376496 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.376507 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.478862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.478907 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.478920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.478937 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.478950 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.581124 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.581162 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.581171 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.581185 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.581195 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.684269 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.684300 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.684308 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.684320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.684329 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.786580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.786623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.786632 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.786645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.786654 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.889283 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.889341 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.889356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.889380 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.889397 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.991657 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.991713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.991726 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.991743 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:42 crc kubenswrapper[4725]: I1202 13:05:42.991756 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:42Z","lastTransitionTime":"2025-12-02T13:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.093517 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.093555 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.093567 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.093582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.093592 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.196395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.196431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.196442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.196485 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.196498 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.267527 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.267679 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.267777 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.267856 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.298841 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.298869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.298877 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.298890 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.298899 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.376190 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.376221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.376229 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.376241 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.376251 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.386830 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:43Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.389788 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.389838 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.389850 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.389867 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.389885 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.399629 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:43Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.402869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.402905 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.402913 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.402928 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.402937 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.413523 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:43Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.416759 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.416794 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.416806 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.416823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.416834 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.427500 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:43Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.430582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.430628 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.430641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.430660 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.430674 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.442876 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:43Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:43 crc kubenswrapper[4725]: E1202 13:05:43.442987 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.444623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.444672 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.444688 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.444711 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.444724 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.547301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.547346 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.547357 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.547378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.547390 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.649321 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.649504 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.649524 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.649557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.649576 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.751649 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.751690 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.751700 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.751717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.751728 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.854255 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.854297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.854313 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.854329 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.854339 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.956863 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.956903 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.956915 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.956930 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:43 crc kubenswrapper[4725]: I1202 13:05:43.956940 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:43Z","lastTransitionTime":"2025-12-02T13:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.059331 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.059379 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.059393 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.059413 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.059437 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.161901 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.161957 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.161973 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.162024 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.162040 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.264593 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.264634 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.264644 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.264658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.264668 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.267808 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.267898 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:44 crc kubenswrapper[4725]: E1202 13:05:44.268006 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:44 crc kubenswrapper[4725]: E1202 13:05:44.268084 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.366984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.367021 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.367072 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.367089 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.367100 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.469052 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.469087 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.469097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.469116 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.469125 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.571905 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.571952 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.571967 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.571985 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.572003 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.674746 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.674781 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.674788 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.674802 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.674811 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.776620 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.776665 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.776678 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.776693 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.776703 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.879075 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.879124 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.879137 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.879153 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.879166 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.981568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.981611 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.981622 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.981638 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:44 crc kubenswrapper[4725]: I1202 13:05:44.981649 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:44Z","lastTransitionTime":"2025-12-02T13:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.083342 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.083379 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.083389 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.083407 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.083418 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.185889 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.185932 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.185941 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.185956 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.185966 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.267369 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.267474 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:45 crc kubenswrapper[4725]: E1202 13:05:45.267521 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:45 crc kubenswrapper[4725]: E1202 13:05:45.267562 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.287984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.288021 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.288030 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.288045 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.288054 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.390046 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.390088 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.390098 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.390113 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.390122 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.492567 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.492607 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.492620 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.492637 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.492650 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.595307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.595351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.595360 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.595378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.595389 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.697855 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.697889 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.697898 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.697911 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.697921 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.800369 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.800406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.800417 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.800434 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.800445 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.902410 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.902481 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.902494 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.902509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:45 crc kubenswrapper[4725]: I1202 13:05:45.902518 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:45Z","lastTransitionTime":"2025-12-02T13:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.005186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.005220 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.005229 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.005244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.005253 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.107435 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.107511 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.107522 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.107539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.107551 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.210050 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.210123 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.210139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.210155 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.210166 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.267966 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.268002 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:46 crc kubenswrapper[4725]: E1202 13:05:46.268152 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:46 crc kubenswrapper[4725]: E1202 13:05:46.268213 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.312528 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.312569 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.312578 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.312594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.312603 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.414424 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.414473 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.414482 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.414494 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.414510 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.516649 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.516695 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.516704 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.516719 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.516729 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.619202 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.619242 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.619251 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.619266 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.619277 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.720794 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.720819 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.720828 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.720840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.720848 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.823141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.823209 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.823227 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.823637 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.823691 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.925977 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.926640 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.926658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.926675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:46 crc kubenswrapper[4725]: I1202 13:05:46.926688 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:46Z","lastTransitionTime":"2025-12-02T13:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.028918 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.028958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.028968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.028982 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.028991 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.131182 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.131217 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.131226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.131239 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.131249 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.233570 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.233612 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.233623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.233641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.233653 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.267615 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.267776 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:47 crc kubenswrapper[4725]: E1202 13:05:47.267942 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:47 crc kubenswrapper[4725]: E1202 13:05:47.268652 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.269918 4725 scope.go:117] "RemoveContainer" containerID="efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.336442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.336917 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.337007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.337083 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.337145 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.439860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.440147 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.440221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.440288 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.440369 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.542388 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.542442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.542451 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.542491 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.542507 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.644567 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.644598 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.644605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.644618 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.644626 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.746836 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.746883 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.746897 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.746921 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.746938 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.849503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.849544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.849557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.849573 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.849585 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.952876 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.952921 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.952932 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.953138 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:47 crc kubenswrapper[4725]: I1202 13:05:47.953150 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:47Z","lastTransitionTime":"2025-12-02T13:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.056711 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.056752 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.056761 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.056780 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.056791 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.163352 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.163429 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.163442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.163504 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.163522 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.265868 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.265922 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.265933 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.265949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.265976 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.267995 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:48 crc kubenswrapper[4725]: E1202 13:05:48.268107 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.268293 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:48 crc kubenswrapper[4725]: E1202 13:05:48.268360 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.367986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.368013 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.368021 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.368033 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.368041 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.470432 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.470521 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.470539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.470565 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.470585 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.572855 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.572887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.572895 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.572908 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.572918 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.675547 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.675590 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.675600 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.675616 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.675647 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.726410 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/2.log" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.729552 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.730025 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.744153 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.754681 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.769373 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.777604 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.777660 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.777675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.777702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.777722 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.783780 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.798638 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.812097 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.821577 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.830072 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.840601 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.855920 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.880425 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.880519 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.880540 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.880591 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.880613 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.884047 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.898343 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.913339 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.925276 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.936421 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.951726 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.974183 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.982561 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.982670 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.982698 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.982736 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.982761 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:48Z","lastTransitionTime":"2025-12-02T13:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:48 crc kubenswrapper[4725]: I1202 13:05:48.991268 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:48Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.011250 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.085086 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.085135 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.085146 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.085169 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.085181 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.189645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.189750 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.189777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.189809 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.189829 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.267146 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:49 crc kubenswrapper[4725]: E1202 13:05:49.267549 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.267159 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:49 crc kubenswrapper[4725]: E1202 13:05:49.267702 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.284285 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.292687 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.292738 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.292775 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.292809 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.292822 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.307378 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.321580 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.333389 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.346476 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.360279 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.380183 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.395079 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.395112 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.395123 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.395140 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.395152 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.395774 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.409413 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.427507 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.439612 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.448921 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.460182 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.474410 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.491769 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.500950 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.500992 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.501004 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.501020 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.501030 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.503964 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.516342 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.526894 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.537561 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.603330 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.603591 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.603859 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.604043 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.604208 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.706597 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.706623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.706631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.706643 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.706651 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.734855 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/3.log" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.736063 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/2.log" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.738837 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" exitCode=1 Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.738884 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.738921 4725 scope.go:117] "RemoveContainer" containerID="efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.742775 4725 scope.go:117] "RemoveContainer" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" Dec 02 13:05:49 crc kubenswrapper[4725]: E1202 13:05:49.742989 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.752372 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.763342 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.773919 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.786023 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.804119 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.809023 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.809256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.809265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.809282 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.809291 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.816380 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.826414 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.836664 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.846210 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.858548 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.868692 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.877834 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.888610 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.896738 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.905104 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.911174 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.911198 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.911206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.911219 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.911228 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:49Z","lastTransitionTime":"2025-12-02T13:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.913944 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.926718 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.949669 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efc8f3f186c1a2d57f070c2720f343ecc0cd7aa7189d6fc175257d69eff38f2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:16Z\\\",\\\"message\\\":\\\"5ec6-479b-8881-93dfa8d30337\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-console/networking-console-plugin_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.246\\\\\\\", Port:9443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:16.013991 6374 services_controller.go:451] Built service openshift-kube-apiserver/apiserver cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-apiserver/apiserver_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Pr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:48Z\\\",\\\"message\\\":\\\"s:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:48.618799 6751 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1202 13:05:48.620845 6751 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 13:05:48.620949 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:49 crc kubenswrapper[4725]: I1202 13:05:49.959234 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:49Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.012993 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.013027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.013035 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.013050 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.013058 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.115991 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.116272 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.116335 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.116421 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.116518 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.219235 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.219277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.219286 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.219302 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.219315 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.267152 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:50 crc kubenswrapper[4725]: E1202 13:05:50.267266 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.267412 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:50 crc kubenswrapper[4725]: E1202 13:05:50.267478 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.320959 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.321002 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.321013 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.321028 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.321038 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.423364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.423401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.423410 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.423424 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.423433 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.525592 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.525642 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.525655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.525676 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.525690 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.627294 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.627330 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.627337 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.627351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.627359 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.729768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.729807 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.729826 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.729842 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.729855 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.744105 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/3.log" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.747299 4725 scope.go:117] "RemoveContainer" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" Dec 02 13:05:50 crc kubenswrapper[4725]: E1202 13:05:50.747434 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.760901 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.775723 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.785746 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.798987 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.811598 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.824028 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.831724 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.831755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.831764 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.831780 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.831790 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.835420 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.849601 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.868208 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:48Z\\\",\\\"message\\\":\\\"s:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:48.618799 6751 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1202 13:05:48.620845 6751 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 13:05:48.620949 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.887056 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.896687 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.909262 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.920372 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.932921 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.934536 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.934568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.934577 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.934594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.934606 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:50Z","lastTransitionTime":"2025-12-02T13:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.944950 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.958694 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.968939 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:50 crc kubenswrapper[4725]: I1202 13:05:50.983539 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.001704 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:50Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.037893 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.037928 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.037935 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.037950 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.037958 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.140813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.140848 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.140857 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.140873 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.140882 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.243301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.243358 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.243373 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.243394 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.243409 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.268121 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.268124 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:51 crc kubenswrapper[4725]: E1202 13:05:51.268255 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:51 crc kubenswrapper[4725]: E1202 13:05:51.268337 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.345865 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.345902 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.345912 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.345927 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.345937 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.448593 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.448629 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.448639 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.448652 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.448660 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.550931 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.550989 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.550998 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.551014 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.551024 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.653413 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.653481 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.653492 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.653509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.653519 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.755638 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.755676 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.755687 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.755702 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.755713 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.857923 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.857951 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.857959 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.857971 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.857980 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.959876 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.959921 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.959939 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.959960 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:51 crc kubenswrapper[4725]: I1202 13:05:51.959972 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:51Z","lastTransitionTime":"2025-12-02T13:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.062362 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.062392 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.062401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.062415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.062425 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.165085 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.165126 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.165136 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.165147 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.165156 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.210840 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.210911 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.210941 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.210960 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.210981 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211077 4725 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211111 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:56.211052431 +0000 UTC m=+147.167694186 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211170 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211184 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211194 4725 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211243 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211253 4725 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211260 4725 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211173 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:06:56.211151173 +0000 UTC m=+147.167793018 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211272 4725 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211295 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 13:06:56.211282576 +0000 UTC m=+147.167924271 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211558 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 13:06:56.211504181 +0000 UTC m=+147.168146006 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.211590 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 13:06:56.211573373 +0000 UTC m=+147.168215298 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267091 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267091 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.267230 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:52 crc kubenswrapper[4725]: E1202 13:05:52.267297 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267609 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267618 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.267639 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.370350 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.370398 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.370412 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.370435 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.370450 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.473066 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.473385 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.473401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.473418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.473429 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.574982 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.575015 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.575025 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.575039 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.575049 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.677368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.677411 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.677428 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.677448 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.677490 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.779506 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.779538 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.779549 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.779564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.779573 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.881415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.881442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.881450 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.881488 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.881506 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.983977 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.984014 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.984022 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.984037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:52 crc kubenswrapper[4725]: I1202 13:05:52.984049 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:52Z","lastTransitionTime":"2025-12-02T13:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.086288 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.086347 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.086363 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.086383 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.086395 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.188604 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.188899 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.189012 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.189164 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.189274 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.267565 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.267565 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.267718 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.267837 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.292067 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.292282 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.292386 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.292504 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.292607 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.394432 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.394491 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.394503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.394516 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.394524 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.497875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.497919 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.497940 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.497960 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.497974 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.600583 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.600632 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.600641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.600655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.600664 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.703041 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.703076 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.703084 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.703098 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.703107 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.716571 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.716619 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.716630 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.716646 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.716657 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.736105 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.739905 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.739933 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.739941 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.739954 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.739963 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.751980 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.756586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.756629 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.756644 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.756665 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.756679 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.767925 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.771869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.771901 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.771910 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.771926 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.771938 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.783349 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.787093 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.787123 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.787134 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.787148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.787157 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.798021 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:53Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:53 crc kubenswrapper[4725]: E1202 13:05:53.798134 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.805134 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.805174 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.805187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.805203 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.805214 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.907910 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.907981 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.908007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.908037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:53 crc kubenswrapper[4725]: I1202 13:05:53.908060 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:53Z","lastTransitionTime":"2025-12-02T13:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.010741 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.010786 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.010798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.010815 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.010826 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.114138 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.114178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.114188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.114206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.114216 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.216504 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.216562 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.216575 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.216594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.216609 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.267147 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.267240 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:54 crc kubenswrapper[4725]: E1202 13:05:54.267286 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:54 crc kubenswrapper[4725]: E1202 13:05:54.267398 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.319395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.319444 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.319478 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.319502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.319514 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.422695 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.422764 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.422777 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.422799 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.422812 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.525144 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.525181 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.525189 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.525201 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.525209 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.627771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.627811 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.627822 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.627839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.627851 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.729939 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.729972 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.729981 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.730004 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.730019 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.832187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.832226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.832234 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.832249 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.832259 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.934583 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.934639 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.934655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.934734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:54 crc kubenswrapper[4725]: I1202 13:05:54.934784 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:54Z","lastTransitionTime":"2025-12-02T13:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.036754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.036819 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.036828 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.036843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.036853 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.138895 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.138932 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.138943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.138958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.138968 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.241286 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.241324 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.241333 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.241347 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.241356 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.268129 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:55 crc kubenswrapper[4725]: E1202 13:05:55.268275 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.268333 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:55 crc kubenswrapper[4725]: E1202 13:05:55.268519 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.343426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.343515 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.343560 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.343582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.343597 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.445378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.445428 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.445445 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.445509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.445533 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.548030 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.548070 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.548084 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.548101 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.548112 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.651069 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.651106 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.651118 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.651134 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.651144 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.753331 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.753392 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.753402 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.753417 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.753428 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.855375 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.855418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.855430 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.855532 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.855546 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.958839 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.958874 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.958886 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.958908 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:55 crc kubenswrapper[4725]: I1202 13:05:55.958922 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:55Z","lastTransitionTime":"2025-12-02T13:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.061095 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.061136 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.061150 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.061167 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.061179 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.163963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.164014 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.164025 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.164041 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.164054 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.266614 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.266650 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.266660 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.266675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.266685 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.267165 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.267269 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:56 crc kubenswrapper[4725]: E1202 13:05:56.267452 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:56 crc kubenswrapper[4725]: E1202 13:05:56.267625 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.369108 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.369147 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.369156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.369171 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.369180 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.471924 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.471980 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.471997 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.472022 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.472040 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.573657 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.573723 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.573751 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.573767 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.573778 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.676190 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.676256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.676267 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.676283 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.676295 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.778963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.779002 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.779012 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.779028 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.779038 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.882268 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.882336 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.882368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.882397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.882417 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.984914 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.984957 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.984968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.984983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:56 crc kubenswrapper[4725]: I1202 13:05:56.984992 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:56Z","lastTransitionTime":"2025-12-02T13:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.088138 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.088179 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.088188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.088204 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.088218 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.191186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.191249 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.191262 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.191280 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.191292 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.267201 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.267248 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:57 crc kubenswrapper[4725]: E1202 13:05:57.267324 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:57 crc kubenswrapper[4725]: E1202 13:05:57.267377 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.293720 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.293762 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.293771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.293788 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.293800 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.395909 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.395958 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.395968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.395983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.395995 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.498574 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.498624 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.498635 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.498654 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.498666 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.601050 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.601092 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.601102 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.601117 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.601128 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.703355 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.703428 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.703450 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.703507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.703533 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.806117 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.806167 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.806179 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.806194 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.806205 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.908421 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.908479 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.908493 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.908510 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:57 crc kubenswrapper[4725]: I1202 13:05:57.908521 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:57Z","lastTransitionTime":"2025-12-02T13:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.011320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.011357 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.011366 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.011379 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.011390 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.114939 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.114985 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.114994 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.115007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.115016 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.217505 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.217545 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.217554 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.217567 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.217576 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.267987 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.268015 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:05:58 crc kubenswrapper[4725]: E1202 13:05:58.268114 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:05:58 crc kubenswrapper[4725]: E1202 13:05:58.268209 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.320094 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.320156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.320168 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.320186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.320197 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.421891 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.421938 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.421950 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.421968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.421980 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.524371 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.524429 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.524439 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.524474 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.524495 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.627418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.627480 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.627488 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.627504 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.627512 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.729808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.729846 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.729857 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.729875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.729888 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.831945 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.831988 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.832000 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.832015 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.832026 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.934556 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.934623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.934636 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.934656 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:58 crc kubenswrapper[4725]: I1202 13:05:58.934670 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:58Z","lastTransitionTime":"2025-12-02T13:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.037980 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.038036 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.038054 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.038078 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.038096 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.140686 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.140726 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.140735 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.140749 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.140759 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.243536 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.243570 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.243580 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.243594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.243604 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.268516 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:05:59 crc kubenswrapper[4725]: E1202 13:05:59.268747 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.268907 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:05:59 crc kubenswrapper[4725]: E1202 13:05:59.269152 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.284548 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.298214 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.310804 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.327095 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.346374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.346426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.346437 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.346476 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.346490 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.351624 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.366338 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.381129 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.402946 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.415853 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.427316 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.440882 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.448887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.449893 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.449920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.449937 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.449954 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.455598 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.473314 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.484156 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.494532 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.507852 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.549794 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.552197 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.552265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.552280 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.552308 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.552326 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.578776 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:48Z\\\",\\\"message\\\":\\\"s:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:48.618799 6751 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1202 13:05:48.620845 6751 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 13:05:48.620949 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.593617 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:05:59Z is after 2025-08-24T17:21:41Z" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.655604 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.655660 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.655675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.655699 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.655712 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.758372 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.758430 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.758440 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.758471 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.758481 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.860880 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.861500 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.861681 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.861875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.862080 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.965502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.965581 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.965598 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.965631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:05:59 crc kubenswrapper[4725]: I1202 13:05:59.965650 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:05:59Z","lastTransitionTime":"2025-12-02T13:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.068875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.068931 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.068942 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.068963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.068975 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.172511 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.173037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.173142 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.173257 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.173378 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.267930 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.268999 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:00 crc kubenswrapper[4725]: E1202 13:06:00.269120 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:00 crc kubenswrapper[4725]: E1202 13:06:00.269266 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.276281 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.276360 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.276378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.276397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.276409 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.379938 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.379981 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.379992 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.380007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.380019 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.483938 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.483986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.483997 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.484017 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.484032 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.587082 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.587126 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.587142 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.587168 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.587185 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.690720 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.690793 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.690820 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.690845 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.690876 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.793608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.793655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.793669 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.793691 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.793707 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.896750 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.896826 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.896847 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.896875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.896898 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.999861 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.999899 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.999908 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.999922 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:00 crc kubenswrapper[4725]: I1202 13:06:00.999931 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:00Z","lastTransitionTime":"2025-12-02T13:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.103290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.103336 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.103350 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.103373 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.103386 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.206290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.206358 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.206368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.206401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.206413 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.267450 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:01 crc kubenswrapper[4725]: E1202 13:06:01.267599 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.267660 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:01 crc kubenswrapper[4725]: E1202 13:06:01.267915 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.309442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.309503 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.309514 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.309530 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.309541 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.412821 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.412882 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.412900 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.412921 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.412937 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.515890 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.515946 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.515964 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.515986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.516003 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.619178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.619272 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.619301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.619376 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.619401 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.723052 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.723161 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.723266 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.723305 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.723327 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.826068 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.826132 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.826150 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.826174 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.826191 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.929323 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.929364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.929378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.929401 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:01 crc kubenswrapper[4725]: I1202 13:06:01.929416 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:01Z","lastTransitionTime":"2025-12-02T13:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.032430 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.032509 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.032525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.032541 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.032553 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.134586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.134658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.134675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.134698 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.134713 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.236605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.236714 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.236775 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.236798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.236818 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.267223 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:02 crc kubenswrapper[4725]: E1202 13:06:02.267341 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.267420 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:02 crc kubenswrapper[4725]: E1202 13:06:02.267727 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.339771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.339808 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.339841 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.339857 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.339869 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.443625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.443684 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.443701 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.443725 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.443750 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.546776 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.546868 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.546885 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.546909 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.546922 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.650331 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.650403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.650423 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.650451 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.650516 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.753712 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.753769 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.753786 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.753813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.753833 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.857630 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.857705 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.857726 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.857757 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.857780 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.962521 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.962600 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.962623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.962658 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:02 crc kubenswrapper[4725]: I1202 13:06:02.962683 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:02Z","lastTransitionTime":"2025-12-02T13:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.066297 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.066361 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.066380 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.066406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.066425 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.169433 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.169542 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.169570 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.169611 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.169653 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.268110 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.268126 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:03 crc kubenswrapper[4725]: E1202 13:06:03.268588 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:03 crc kubenswrapper[4725]: E1202 13:06:03.268749 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.272088 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.272127 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.272139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.272156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.272168 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.375949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.375983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.375991 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.376004 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.376014 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.477949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.477994 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.478012 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.478030 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.478041 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.581145 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.581193 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.581203 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.581221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.581231 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.684165 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.684226 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.684239 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.684259 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.684272 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.787309 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.787356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.787367 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.787382 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.787393 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.891432 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.891560 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.891586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.891622 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.891669 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.932519 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.932713 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.932807 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.932900 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.933003 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: E1202 13:06:03.952250 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:03Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.959366 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.959406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.959415 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.959431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.959442 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:03 crc kubenswrapper[4725]: E1202 13:06:03.974981 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:03Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.981064 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.981111 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.981125 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.981141 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:03 crc kubenswrapper[4725]: I1202 13:06:03.981151 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:03Z","lastTransitionTime":"2025-12-02T13:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: E1202 13:06:04.002570 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:03Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.007799 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.007869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.007886 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.007908 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.007921 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: E1202 13:06:04.025730 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:04Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.030413 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.030526 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.030544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.030593 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.030607 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: E1202 13:06:04.045822 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:04Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:04 crc kubenswrapper[4725]: E1202 13:06:04.046004 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.048434 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.048516 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.048532 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.048573 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.048586 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.151977 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.152049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.152064 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.152090 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.152106 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.255767 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.255845 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.255866 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.255895 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.255930 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.267238 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.267579 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:04 crc kubenswrapper[4725]: E1202 13:06:04.267773 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:04 crc kubenswrapper[4725]: E1202 13:06:04.267983 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.359619 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.359735 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.359750 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.359775 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.359789 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.463446 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.463533 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.463548 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.463568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.463583 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.566277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.566346 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.566369 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.566396 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.566417 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.669031 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.669073 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.669083 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.669097 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.669108 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.772627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.772680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.772694 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.772717 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.772795 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.875512 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.875556 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.875568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.875587 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.875599 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.978148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.978185 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.978193 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.978207 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:04 crc kubenswrapper[4725]: I1202 13:06:04.978218 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:04Z","lastTransitionTime":"2025-12-02T13:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.081245 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.081287 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.081296 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.081311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.081322 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.183644 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.183729 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.183769 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.183809 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.183837 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.267255 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:05 crc kubenswrapper[4725]: E1202 13:06:05.267414 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.267609 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:05 crc kubenswrapper[4725]: E1202 13:06:05.267976 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.286309 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.286394 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.286407 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.286425 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.286437 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.388979 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.389031 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.389043 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.389063 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.389077 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.491857 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.491917 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.491929 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.491943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.491953 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.594780 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.594831 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.594843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.594862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.594875 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.697784 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.697833 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.697847 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.697867 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.697881 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.800169 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.800225 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.800240 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.800262 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.800277 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.902890 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.902932 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.902943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.902959 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:05 crc kubenswrapper[4725]: I1202 13:06:05.902970 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:05Z","lastTransitionTime":"2025-12-02T13:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.005740 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.005793 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.005802 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.005815 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.005826 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.108666 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.108739 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.108757 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.108783 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.108801 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.212006 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.212270 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.212403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.212523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.212631 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.267651 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:06 crc kubenswrapper[4725]: E1202 13:06:06.267751 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.267973 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:06 crc kubenswrapper[4725]: E1202 13:06:06.268621 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.270259 4725 scope.go:117] "RemoveContainer" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" Dec 02 13:06:06 crc kubenswrapper[4725]: E1202 13:06:06.270749 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.316514 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.316575 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.316587 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.316625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.316636 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.361862 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:06 crc kubenswrapper[4725]: E1202 13:06:06.362280 4725 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:06:06 crc kubenswrapper[4725]: E1202 13:06:06.362378 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs podName:9feadd7b-f199-41c3-b908-43774e1cd2f3 nodeName:}" failed. No retries permitted until 2025-12-02 13:07:10.362360198 +0000 UTC m=+161.319001893 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs") pod "network-metrics-daemon-vs5fz" (UID: "9feadd7b-f199-41c3-b908-43774e1cd2f3") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.419440 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.419525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.419539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.419557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.419568 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.522592 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.522636 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.522651 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.522667 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.522681 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.626289 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.626345 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.626364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.626393 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.626417 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.729439 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.729553 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.729572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.729590 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.729604 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.831418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.831486 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.831505 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.831523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.831535 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.934056 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.934103 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.934116 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.934135 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:06 crc kubenswrapper[4725]: I1202 13:06:06.934148 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:06Z","lastTransitionTime":"2025-12-02T13:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.036200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.036242 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.036256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.036274 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.036287 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.139582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.139641 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.139651 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.139674 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.139699 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.243153 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.243277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.243293 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.243315 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.243337 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.267329 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.267365 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:07 crc kubenswrapper[4725]: E1202 13:06:07.267626 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:07 crc kubenswrapper[4725]: E1202 13:06:07.267821 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.345861 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.345895 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.345904 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.345917 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.345925 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.449326 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.449390 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.449407 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.449431 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.449448 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.552199 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.552301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.552320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.552354 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.552374 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.655034 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.655107 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.655131 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.655161 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.655186 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.757268 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.757315 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.757327 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.757343 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.757353 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.859955 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.859996 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.860007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.860025 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.860038 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.962270 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.962311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.962322 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.962336 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:07 crc kubenswrapper[4725]: I1202 13:06:07.962347 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:07Z","lastTransitionTime":"2025-12-02T13:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.064948 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.064997 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.065006 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.065023 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.065037 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.166798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.166832 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.166840 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.166853 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.166878 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.267752 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.267756 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:08 crc kubenswrapper[4725]: E1202 13:06:08.267887 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:08 crc kubenswrapper[4725]: E1202 13:06:08.267942 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.269243 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.269267 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.269277 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.269289 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.269301 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.371798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.371849 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.371862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.371879 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.371897 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.474983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.475037 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.475053 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.475074 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.475089 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.578162 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.578229 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.578244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.578268 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.578282 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.681290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.681345 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.681365 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.681387 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.681403 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.784597 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.784647 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.784659 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.784675 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.784687 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.888069 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.888150 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.888169 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.888194 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.888212 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.991065 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.991102 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.991119 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.991135 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:08 crc kubenswrapper[4725]: I1202 13:06:08.991145 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:08Z","lastTransitionTime":"2025-12-02T13:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.093098 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.093140 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.093151 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.093167 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.093178 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.196558 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.196642 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.196664 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.196691 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.196714 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.268140 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.268261 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:09 crc kubenswrapper[4725]: E1202 13:06:09.268339 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:09 crc kubenswrapper[4725]: E1202 13:06:09.268417 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.288016 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fab718d-f1fa-42a2-bf76-177639175702\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f248c9d4f05d9e4ba65f3cf45f84842887a7b57fcac9f6487c7ca4d09bc5cfac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b5f4ca34bb7e51bb912421febb5697b4ea288772179d4e16be58dba82c10fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.299708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.299771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.299790 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.299814 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.299830 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.307373 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.327968 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab1ce6ac-3a16-49c5-b3db-6ca391e3aff5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://777de3ac1b7ba004ae8bbe4d63989caa9f3aa87ee309f57b90f746b3668fc76d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2fd4c42d416793e08f8a500b99debbc45412777739ef458538c466265b75bf56\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e598b9ef662703bcf5b7b8a78f7685f42834c7880601a71da0cbc1d9bc0a9226\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://945efa6370377bc9a13c701afd580c013d5a0058c5fde8524bfa9e3706a3c276\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd4daf1704d365efba6ddaf2c54a03f3ec147f5f87cf6f087af9a46a88530404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a742fc5027dd0fc60b8ee5b3c8ab4bea1ff4c662b36c6116c5107bbb75f19f44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0896cd12766e4d484c4dd06be0e3c101aca99faf2af3c26829f208134fa94784\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbpks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xcgm9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.358497 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f939ed3f-9402-4a57-858f-0323084742a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:48Z\\\",\\\"message\\\":\\\"s:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 13:05:48.618799 6751 ovn.go:134] Ensuring zone local for Pod openshift-etcd/etcd-crc in node crc\\\\nI1202 13:05:48.620845 6751 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-controller-manager/kube-controller-manager]} name:Service_openshift-kube-controller-manager/kube-controller-manager_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ba175bbe-5cc4-47e6-a32d-57693e1320bd}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 13:05:48.620949 6751 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:05:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-btq9p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-nnqsk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.378874 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9feadd7b-f199-41c3-b908-43774e1cd2f3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbhzm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:02Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vs5fz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.398291 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.402451 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.402523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.402539 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.402561 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.402577 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.411662 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://279a6c1b3cbb15e880b54c08ee97695a99a83b40c6e59c061ff623ad68544549\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.425948 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c50af795-4a0f-45cd-b117-f6ddea79ee51\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://316b6750f5223502dac207a38c528a62399f1f177b9c324d0e3824d4625febf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l4gzp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zc2vm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.443706 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5b3cacb-78af-4fb7-a149-9d500e22828f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T13:04:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1202 13:04:47.731033 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1202 13:04:47.731230 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 13:04:47.732356 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3279971628/tls.crt::/tmp/serving-cert-3279971628/tls.key\\\\\\\"\\\\nI1202 13:04:47.962734 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 13:04:47.964761 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 13:04:47.964779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 13:04:47.964799 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 13:04:47.964813 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 13:04:47.968894 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1202 13:04:47.968907 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1202 13:04:47.968953 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968960 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 13:04:47.968965 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 13:04:47.968969 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 13:04:47.968973 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 13:04:47.968977 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1202 13:04:47.971147 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.462781 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"82b45ae7-c8f4-4402-ac8a-477d6909c10f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ab467c94262f5da4f372170a5e9d4f353df6432efd6cdfd324806130372b7ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f062e2b5efd90b36f35dfb8b2d5137d1b5a17925105c19991c475c7a8598088\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08f95fca4fae3776b9165837badf27d39304e5ee1169066a32894d9cd84e411f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://586e678e6ee7674c58c9c4b9c384ac1ea867db84b65ec6d819daad9f253fc090\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dc1de3f7fb2b07238a91e8c7bb15b94b5b66c43bd787180d868633124e9801\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://abc69209714bcb932c56c0de7b5bfcbe5e435962bebe61ad196f116ffeae709a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e68c407ec35644916ba8ee9890a2527a8a0df5dd7722b9c4053fec7f6a33f08d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e7dfdbaa6c2283a1facab35459b5b2a0752eb10dc4f8f84bc206742015dabda\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.479309 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b121a5cc963ecf62d64c43469034842dd6e4f9609cbe6937fed7454bffbdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://935f7200c65460c96ad65eac0b5fbd3e6cd34bb8c95ce69a8b7c5c1aa51bd825\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.492326 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e61071d0-15cb-43d9-a3f1-df802d821e1f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c8ed3390eff1ab81b1bbbc3bcf0c4dbde4a9799383b5c2e3a294d5e407c5879\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da747b62362184bf64a1f8ec610c5f7125a23489d68714c1cd777670710f58e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pwnlt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:05:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tdvqh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.504799 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.505579 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.505617 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.505630 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.505647 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.505658 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.516433 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.528631 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.542277 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.551269 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5zhgg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45c29b81-1929-415a-a671-606bae374a50\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a55c9740ea59411e29b711848f954bfd16f1e4b93502620d998d6c43dafb85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvjhc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5zhgg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.563325 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8s8qq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T13:05:34Z\\\",\\\"message\\\":\\\"2025-12-02T13:04:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041\\\\n2025-12-02T13:04:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_3cb6ad27-c8f6-48d5-848f-060965308041 to /host/opt/cni/bin/\\\\n2025-12-02T13:04:49Z [verbose] multus-daemon started\\\\n2025-12-02T13:04:49Z [verbose] Readiness Indicator file check\\\\n2025-12-02T13:05:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:05:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8fz4d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:48Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8s8qq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.572778 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q658f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e0803379-a77d-4799-a5ef-b2cfe94949a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1f26a24a77115e6bf707eb78c14a13848c12a2f3986aac216fea478d3df26d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cvjcq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q658f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:09Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.608162 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.608215 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.608228 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.608251 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.608262 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.710368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.710404 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.710413 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.710426 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.710434 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.813311 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.813348 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.813357 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.813371 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.813380 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.917230 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.917281 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.917290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.917308 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:09 crc kubenswrapper[4725]: I1202 13:06:09.917322 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:09Z","lastTransitionTime":"2025-12-02T13:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.019981 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.020033 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.020071 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.020094 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.020105 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.122696 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.122745 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.122754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.122778 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.122789 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.225488 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.225519 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.225527 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.225543 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.225586 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.267564 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.267625 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:10 crc kubenswrapper[4725]: E1202 13:06:10.267707 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:10 crc kubenswrapper[4725]: E1202 13:06:10.267777 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.328448 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.328617 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.328642 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.328667 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.328684 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.431138 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.431178 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.431193 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.431209 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.431222 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.534295 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.534349 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.534359 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.534374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.534384 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.637397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.637436 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.637444 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.637478 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.637488 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.740148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.740244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.740264 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.740294 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.740314 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.842829 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.842870 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.842879 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.842892 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.842902 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.944900 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.944966 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.944989 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.945017 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:10 crc kubenswrapper[4725]: I1202 13:06:10.945036 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:10Z","lastTransitionTime":"2025-12-02T13:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.047918 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.047963 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.047972 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.047988 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.047998 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.151220 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.151279 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.151290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.151312 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.151324 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.253934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.253972 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.253985 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.254002 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.254013 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.267649 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:11 crc kubenswrapper[4725]: E1202 13:06:11.267899 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.268030 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:11 crc kubenswrapper[4725]: E1202 13:06:11.268197 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.356968 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.357016 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.357027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.357047 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.357060 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.460625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.460672 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.460680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.460697 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.460706 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.563568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.563613 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.563622 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.563637 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.563645 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.667219 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.667287 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.667310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.667339 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.667358 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.770812 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.770858 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.770869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.770886 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.770918 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.874307 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.874382 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.874397 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.874420 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.874438 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.978291 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.978375 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.978400 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.978442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:11 crc kubenswrapper[4725]: I1202 13:06:11.978506 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:11Z","lastTransitionTime":"2025-12-02T13:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.082940 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.082986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.082997 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.083017 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.083027 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.185122 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.185219 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.185246 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.185279 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.185298 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.268075 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.268096 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:12 crc kubenswrapper[4725]: E1202 13:06:12.268341 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:12 crc kubenswrapper[4725]: E1202 13:06:12.268423 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.288345 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.288395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.288407 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.288425 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.288439 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.391476 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.391505 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.391513 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.391528 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.391538 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.494669 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.494728 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.494741 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.494762 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.494775 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.597004 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.597062 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.597079 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.597100 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.597115 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.699538 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.699565 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.699573 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.699589 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.699598 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.802139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.802199 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.802208 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.802221 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.802229 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.904689 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.904730 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.904740 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.904754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:12 crc kubenswrapper[4725]: I1202 13:06:12.904763 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:12Z","lastTransitionTime":"2025-12-02T13:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.007887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.007961 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.007980 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.008010 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.008040 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.110943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.110990 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.111008 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.111027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.111038 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.214272 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.214349 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.214368 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.214402 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.214424 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.267781 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.267802 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:13 crc kubenswrapper[4725]: E1202 13:06:13.268092 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:13 crc kubenswrapper[4725]: E1202 13:06:13.268538 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.317621 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.317662 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.317677 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.317697 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.317711 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.420649 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.420708 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.420722 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.420741 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.420751 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.524632 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.524722 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.524734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.524751 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.524767 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.628419 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.628566 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.628586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.628618 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.628639 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.731525 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.731594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.731609 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.731625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.731637 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.835271 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.835332 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.835343 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.835406 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.835417 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.938230 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.938273 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.938284 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.938299 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:13 crc kubenswrapper[4725]: I1202 13:06:13.938310 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:13Z","lastTransitionTime":"2025-12-02T13:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.040257 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.040301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.040313 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.040330 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.040341 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.142686 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.142735 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.142745 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.142761 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.142771 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.245813 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.245852 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.245862 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.245876 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.245885 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.267367 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.267367 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.267509 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.267540 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.349414 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.349520 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.349544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.349576 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.349598 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.396935 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.397006 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.397029 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.397059 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.397081 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.414047 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:14Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.418743 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.418771 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.418782 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.418820 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.418835 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.432782 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:14Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.437596 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.437642 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.437655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.437673 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.437689 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.452036 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:14Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.456682 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.456733 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.456751 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.456770 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.456784 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.471035 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:14Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.476212 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.476271 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.476284 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.476303 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.476315 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.493878 4725 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T13:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6a5556-c77a-40a9-a29a-40e3b11740a7\\\",\\\"systemUUID\\\":\\\"0121cc7c-04ba-4d57-99ad-07680ad4b3d9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:14Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:14 crc kubenswrapper[4725]: E1202 13:06:14.494053 4725 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.496064 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.496117 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.496131 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.496153 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.496167 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.599950 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.600265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.600341 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.600449 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.600563 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.702980 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.703057 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.703076 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.703096 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.703107 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.805603 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.805645 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.805657 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.805673 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.805683 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.907523 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.907596 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.907608 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.907624 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:14 crc kubenswrapper[4725]: I1202 13:06:14.907635 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:14Z","lastTransitionTime":"2025-12-02T13:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.009364 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.009444 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.009489 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.009506 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.009516 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.111080 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.111116 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.111125 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.111138 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.111146 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.213814 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.213882 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.213903 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.213918 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.213928 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.267713 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:15 crc kubenswrapper[4725]: E1202 13:06:15.267974 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.267750 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:15 crc kubenswrapper[4725]: E1202 13:06:15.268363 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.317543 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.317887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.317986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.318091 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.318188 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.421987 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.422274 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.422408 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.422538 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.422626 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.525132 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.525187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.525197 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.525214 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.525222 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.629507 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.629575 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.629592 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.629615 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.629629 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.733549 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.733614 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.733636 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.733681 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.733712 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.837373 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.837438 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.837454 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.837497 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.837511 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.940934 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.941216 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.941302 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.941391 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:15 crc kubenswrapper[4725]: I1202 13:06:15.941514 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:15Z","lastTransitionTime":"2025-12-02T13:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.045491 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.045555 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.045590 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.045635 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.045659 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.148908 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.149605 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.149630 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.149659 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.149675 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.252026 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.252058 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.252069 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.252086 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.252097 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.267961 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:16 crc kubenswrapper[4725]: E1202 13:06:16.268081 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.268614 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:16 crc kubenswrapper[4725]: E1202 13:06:16.268871 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.353943 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.353983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.353995 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.354015 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.354028 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.456152 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.456187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.456195 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.456208 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.456219 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.558442 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.558534 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.558545 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.558563 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.558575 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.660820 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.660859 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.660869 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.660887 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.660898 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.763679 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.763787 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.763854 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.763878 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.763924 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.866143 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.866186 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.866210 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.866231 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.866246 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.968693 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.968760 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.968787 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.968823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:16 crc kubenswrapper[4725]: I1202 13:06:16.968847 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:16Z","lastTransitionTime":"2025-12-02T13:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.070496 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.070543 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.070554 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.070568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.070578 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.173121 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.173158 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.173171 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.173187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.173198 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.268382 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:17 crc kubenswrapper[4725]: E1202 13:06:17.268481 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.268609 4725 scope.go:117] "RemoveContainer" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.268629 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:17 crc kubenswrapper[4725]: E1202 13:06:17.268699 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:17 crc kubenswrapper[4725]: E1202 13:06:17.268729 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-nnqsk_openshift-ovn-kubernetes(f939ed3f-9402-4a57-858f-0323084742a9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.274866 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.274903 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.274949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.274965 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.274976 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.377260 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.377294 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.377310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.377326 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.377336 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.479103 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.479139 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.479149 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.479165 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.479175 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.581763 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.581804 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.581816 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.581832 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.581842 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.684773 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.684835 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.684846 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.684858 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.684866 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.787330 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.787384 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.787398 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.787414 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.787423 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.889582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.889624 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.889638 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.889653 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.889664 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.991800 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.991844 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.991860 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.991881 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:17 crc kubenswrapper[4725]: I1202 13:06:17.991893 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:17Z","lastTransitionTime":"2025-12-02T13:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.094275 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.094312 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.094321 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.094334 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.094342 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.196200 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.196258 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.196268 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.196283 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.196293 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.267918 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.268113 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:18 crc kubenswrapper[4725]: E1202 13:06:18.268262 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:18 crc kubenswrapper[4725]: E1202 13:06:18.268431 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.298754 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.298790 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.298798 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.298811 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.298820 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.401052 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.401086 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.401095 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.401109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.401120 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.503383 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.503410 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.503418 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.503432 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.503441 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.606111 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.606148 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.606156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.606170 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.606180 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.709111 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.709167 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.709180 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.709199 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.709211 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.812301 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.812348 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.812359 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.812374 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.812387 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.914880 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.914983 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.915009 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.915046 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:18 crc kubenswrapper[4725]: I1202 13:06:18.915071 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:18Z","lastTransitionTime":"2025-12-02T13:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.017674 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.017720 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.017731 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.017748 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.017760 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.119973 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.120022 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.120034 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.120054 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.120066 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.222205 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.222255 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.222265 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.222290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.222301 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.267548 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.267653 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:19 crc kubenswrapper[4725]: E1202 13:06:19.267870 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:19 crc kubenswrapper[4725]: E1202 13:06:19.268048 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.279103 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9576f4cb-b74d-48db-a6f8-2b3288d0b54b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e70e1d126c65aad51f3e2d37d1ba24f17a602167f9d71474fa44a0ec56871c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aafdd5723fa9d88c4eca16dc11b4932e22428bd25c68692fa43f873485a9fd11\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa4e95ea2e8d1b765f12796f907041f925287434fcd264dd10513268784f1004\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.289853 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2291ea56-7067-4950-8e35-d5795ca0d8db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:05:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7bf1862ab81ccf32ae65c309892a4e76f1d10d836daa7f331c9c4c218fbfbe0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8908982fc3cfa2500fd576da57d030f3c79242b250160809e80d18685686c0fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e85b4d89c540bff0888a7bfa23c29f1982d9e39ac3839873734f262eb8c4637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://633030487ecd1850acc921a6bb4dd17e0ba98b78c8cd23ec4f2ff62432a6bcf2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T13:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T13:04:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T13:04:29Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.300264 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://31a0f2cd1f76dfd6bef4a1cc1006eff41777d04580d820aee3f30ed6451051e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T13:04:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.310699 4725 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T13:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T13:06:19Z is after 2025-08-24T17:21:41Z" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.324266 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.324303 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.324313 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.324328 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.324337 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.339326 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5zhgg" podStartSLOduration=91.3393042 podStartE2EDuration="1m31.3393042s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.325490497 +0000 UTC m=+110.282132192" watchObservedRunningTime="2025-12-02 13:06:19.3393042 +0000 UTC m=+110.295945895" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.350744 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8s8qq" podStartSLOduration=91.350727036 podStartE2EDuration="1m31.350727036s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.339847294 +0000 UTC m=+110.296488999" watchObservedRunningTime="2025-12-02 13:06:19.350727036 +0000 UTC m=+110.307368731" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.360441 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-q658f" podStartSLOduration=91.36042289 podStartE2EDuration="1m31.36042289s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.350542052 +0000 UTC m=+110.307183737" watchObservedRunningTime="2025-12-02 13:06:19.36042289 +0000 UTC m=+110.317064585" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.371383 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=43.371362444 podStartE2EDuration="43.371362444s" podCreationTimestamp="2025-12-02 13:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.36083056 +0000 UTC m=+110.317472255" watchObservedRunningTime="2025-12-02 13:06:19.371362444 +0000 UTC m=+110.328004139" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.395146 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xcgm9" podStartSLOduration=91.395094627 podStartE2EDuration="1m31.395094627s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.395039175 +0000 UTC m=+110.351680880" watchObservedRunningTime="2025-12-02 13:06:19.395094627 +0000 UTC m=+110.351736322" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.425843 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.425885 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.425898 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.425916 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.425927 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.467814 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podStartSLOduration=91.467793701 podStartE2EDuration="1m31.467793701s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.4673436 +0000 UTC m=+110.423985285" watchObservedRunningTime="2025-12-02 13:06:19.467793701 +0000 UTC m=+110.424435396" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.482820 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.482803133 podStartE2EDuration="1m31.482803133s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.482800393 +0000 UTC m=+110.439442098" watchObservedRunningTime="2025-12-02 13:06:19.482803133 +0000 UTC m=+110.439444828" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.510187 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=90.510165693 podStartE2EDuration="1m30.510165693s" podCreationTimestamp="2025-12-02 13:04:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.508203956 +0000 UTC m=+110.464845681" watchObservedRunningTime="2025-12-02 13:06:19.510165693 +0000 UTC m=+110.466807388" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.532247 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.532280 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.532290 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.532302 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.532311 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.635797 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.635831 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.635842 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.635858 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.635869 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.737986 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.738027 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.738035 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.738049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.738059 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.839568 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.839598 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.839609 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.839623 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.839633 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.942261 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.942319 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.942332 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.942351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:19 crc kubenswrapper[4725]: I1202 13:06:19.942363 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:19Z","lastTransitionTime":"2025-12-02T13:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.044892 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.044928 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.044937 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.044951 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.044959 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.146973 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.147021 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.147030 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.147044 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.147052 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.249851 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.249882 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.249891 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.249906 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.249917 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.268182 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:20 crc kubenswrapper[4725]: E1202 13:06:20.268320 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.268632 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:20 crc kubenswrapper[4725]: E1202 13:06:20.268770 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.352196 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.352235 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.352244 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.352296 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.352306 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.454906 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.454949 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.454961 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.454976 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.454987 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.557532 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.558419 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.558479 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.558496 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.558507 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.661420 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.661483 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.661502 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.661520 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.661531 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.764110 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.764140 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.764149 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.764161 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.764169 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.866650 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.866716 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.866734 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.866758 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.866775 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.969267 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.969310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.969319 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.969333 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:20 crc kubenswrapper[4725]: I1202 13:06:20.969342 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:20Z","lastTransitionTime":"2025-12-02T13:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.071625 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.071655 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.071666 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.071680 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.071694 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.174140 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.174176 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.174187 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.174206 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.174214 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.268131 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.268209 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:21 crc kubenswrapper[4725]: E1202 13:06:21.268265 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:21 crc kubenswrapper[4725]: E1202 13:06:21.268427 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.277320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.277377 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.277390 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.277405 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.277418 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.380515 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.380558 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.380567 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.380582 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.380596 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.483494 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.483543 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.483570 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.483594 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.483609 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.585759 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.585795 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.585807 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.585823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.585835 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.687984 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.688033 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.688044 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.688062 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.688075 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.790310 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.790349 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.790361 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.790378 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.790389 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.840960 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/1.log" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.841316 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/0.log" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.841361 4725 generic.go:334] "Generic (PLEG): container finished" podID="4a8c02a6-36ad-4a9c-88b0-064dcc9b7327" containerID="9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913" exitCode=1 Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.841391 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerDied","Data":"9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.841428 4725 scope.go:117] "RemoveContainer" containerID="7a75ec6a749c86a39d6176e1f2a4c5d77821a5b9c1ddc6f26d3e45cf68a4b463" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.841856 4725 scope.go:117] "RemoveContainer" containerID="9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913" Dec 02 13:06:21 crc kubenswrapper[4725]: E1202 13:06:21.842049 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8s8qq_openshift-multus(4a8c02a6-36ad-4a9c-88b0-064dcc9b7327)\"" pod="openshift-multus/multus-8s8qq" podUID="4a8c02a6-36ad-4a9c-88b0-064dcc9b7327" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.855740 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tdvqh" podStartSLOduration=93.855721287 podStartE2EDuration="1m33.855721287s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:19.530954825 +0000 UTC m=+110.487596520" watchObservedRunningTime="2025-12-02 13:06:21.855721287 +0000 UTC m=+112.812362982" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.869993 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=91.869975391 podStartE2EDuration="1m31.869975391s" podCreationTimestamp="2025-12-02 13:04:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:21.856028695 +0000 UTC m=+112.812670390" watchObservedRunningTime="2025-12-02 13:06:21.869975391 +0000 UTC m=+112.826617096" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.870192 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=63.870187146 podStartE2EDuration="1m3.870187146s" podCreationTimestamp="2025-12-02 13:05:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:21.870145295 +0000 UTC m=+112.826786990" watchObservedRunningTime="2025-12-02 13:06:21.870187146 +0000 UTC m=+112.826828851" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.893519 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.893562 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.893572 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.893586 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.893598 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.995768 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.995812 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.995823 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.995841 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:21 crc kubenswrapper[4725]: I1202 13:06:21.995851 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:21Z","lastTransitionTime":"2025-12-02T13:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.098252 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.098299 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.098313 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.098333 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.098345 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.200003 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.200249 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.200327 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.200420 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.200526 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.267928 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:22 crc kubenswrapper[4725]: E1202 13:06:22.268035 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.268200 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:22 crc kubenswrapper[4725]: E1202 13:06:22.268254 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.302557 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.302592 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.302606 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.302627 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.302638 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.405673 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.405744 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.405764 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.405789 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.405805 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.509108 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.509395 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.509486 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.509597 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.509681 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.612036 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.612074 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.612083 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.612096 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.612107 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.714880 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.714920 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.714930 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.714945 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.714955 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.817214 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.817245 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.817256 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.817269 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.817279 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.844937 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/1.log" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.919452 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.919513 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.919528 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.919544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:22 crc kubenswrapper[4725]: I1202 13:06:22.919555 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:22Z","lastTransitionTime":"2025-12-02T13:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.021615 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.021926 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.022011 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.022109 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.022195 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.125320 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.125405 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.125425 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.125449 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.125486 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.226972 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.227019 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.227030 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.227047 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.227060 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.267976 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.268022 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:23 crc kubenswrapper[4725]: E1202 13:06:23.268128 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:23 crc kubenswrapper[4725]: E1202 13:06:23.268284 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.329819 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.329863 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.329875 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.329893 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.329908 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.432229 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.432785 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.432858 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.432927 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.432990 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.535018 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.535049 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.535060 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.535076 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.535086 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.637357 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.637391 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.637403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.637419 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.637429 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.739755 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.739803 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.739814 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.739853 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.739866 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.842119 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.842149 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.842156 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.842169 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.842178 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.945318 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.945351 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.945360 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.945373 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:23 crc kubenswrapper[4725]: I1202 13:06:23.945383 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:23Z","lastTransitionTime":"2025-12-02T13:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.047672 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.048053 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.048266 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.048356 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.048441 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.150906 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.150957 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.150971 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.150991 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.151006 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.253046 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.253305 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.253403 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.253544 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.253688 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.267564 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.267564 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:24 crc kubenswrapper[4725]: E1202 13:06:24.267913 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:24 crc kubenswrapper[4725]: E1202 13:06:24.267851 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.355746 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.356007 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.356092 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.356177 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.356294 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.458951 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.459447 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.459564 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.459631 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.459701 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.562389 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.562433 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.562445 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.562485 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.562498 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.665127 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.665167 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.665179 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.665195 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.665208 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.673131 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.673188 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.673203 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.673280 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.673300 4725 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T13:06:24Z","lastTransitionTime":"2025-12-02T13:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.719627 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl"] Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.719958 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.721980 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.722025 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.722078 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.722406 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.764206 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/777c861a-bf57-4617-b3b2-5305598782b6-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.764806 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/777c861a-bf57-4617-b3b2-5305598782b6-service-ca\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.765024 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/777c861a-bf57-4617-b3b2-5305598782b6-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.765441 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/777c861a-bf57-4617-b3b2-5305598782b6-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.765618 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/777c861a-bf57-4617-b3b2-5305598782b6-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866108 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/777c861a-bf57-4617-b3b2-5305598782b6-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866336 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/777c861a-bf57-4617-b3b2-5305598782b6-service-ca\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866363 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/777c861a-bf57-4617-b3b2-5305598782b6-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866387 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/777c861a-bf57-4617-b3b2-5305598782b6-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866393 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/777c861a-bf57-4617-b3b2-5305598782b6-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866416 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/777c861a-bf57-4617-b3b2-5305598782b6-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.866590 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/777c861a-bf57-4617-b3b2-5305598782b6-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.867651 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/777c861a-bf57-4617-b3b2-5305598782b6-service-ca\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.872293 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/777c861a-bf57-4617-b3b2-5305598782b6-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:24 crc kubenswrapper[4725]: I1202 13:06:24.883558 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/777c861a-bf57-4617-b3b2-5305598782b6-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-n4hnl\" (UID: \"777c861a-bf57-4617-b3b2-5305598782b6\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:25 crc kubenswrapper[4725]: I1202 13:06:25.035621 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" Dec 02 13:06:25 crc kubenswrapper[4725]: I1202 13:06:25.267897 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:25 crc kubenswrapper[4725]: E1202 13:06:25.268521 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:25 crc kubenswrapper[4725]: I1202 13:06:25.268526 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:25 crc kubenswrapper[4725]: E1202 13:06:25.268627 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:25 crc kubenswrapper[4725]: I1202 13:06:25.854900 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" event={"ID":"777c861a-bf57-4617-b3b2-5305598782b6","Type":"ContainerStarted","Data":"f443ff7a1ada003b72366dbdb1b366ccde5ecf94ff2727d7ef486dad066b2e70"} Dec 02 13:06:25 crc kubenswrapper[4725]: I1202 13:06:25.854956 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" event={"ID":"777c861a-bf57-4617-b3b2-5305598782b6","Type":"ContainerStarted","Data":"0c953c3d3df7e59043ab9ecb26808a4c783e2b393c6082b1ae26a4803d3ce06a"} Dec 02 13:06:26 crc kubenswrapper[4725]: I1202 13:06:26.267979 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:26 crc kubenswrapper[4725]: I1202 13:06:26.267988 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:26 crc kubenswrapper[4725]: E1202 13:06:26.268100 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:26 crc kubenswrapper[4725]: E1202 13:06:26.268192 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:27 crc kubenswrapper[4725]: I1202 13:06:27.267607 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:27 crc kubenswrapper[4725]: I1202 13:06:27.267715 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:27 crc kubenswrapper[4725]: E1202 13:06:27.267769 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:27 crc kubenswrapper[4725]: E1202 13:06:27.267876 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:28 crc kubenswrapper[4725]: I1202 13:06:28.267717 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:28 crc kubenswrapper[4725]: I1202 13:06:28.267717 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:28 crc kubenswrapper[4725]: E1202 13:06:28.267832 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:28 crc kubenswrapper[4725]: E1202 13:06:28.267906 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:29 crc kubenswrapper[4725]: I1202 13:06:29.269134 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:29 crc kubenswrapper[4725]: E1202 13:06:29.269269 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:29 crc kubenswrapper[4725]: I1202 13:06:29.269585 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:29 crc kubenswrapper[4725]: E1202 13:06:29.269800 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:29 crc kubenswrapper[4725]: E1202 13:06:29.280197 4725 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 02 13:06:29 crc kubenswrapper[4725]: E1202 13:06:29.391074 4725 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.267915 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.267988 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:30 crc kubenswrapper[4725]: E1202 13:06:30.268188 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:30 crc kubenswrapper[4725]: E1202 13:06:30.268298 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.268874 4725 scope.go:117] "RemoveContainer" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.870845 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/3.log" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.874195 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerStarted","Data":"5f43d87f4c9f0d0bf3882ecef178e296d0b9d78e650fcf8243c0af4028c4a8b6"} Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.874568 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.899901 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podStartSLOduration=102.899885655 podStartE2EDuration="1m42.899885655s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:30.898797699 +0000 UTC m=+121.855439394" watchObservedRunningTime="2025-12-02 13:06:30.899885655 +0000 UTC m=+121.856527350" Dec 02 13:06:30 crc kubenswrapper[4725]: I1202 13:06:30.900346 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-n4hnl" podStartSLOduration=102.900339446 podStartE2EDuration="1m42.900339446s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:25.867420851 +0000 UTC m=+116.824062546" watchObservedRunningTime="2025-12-02 13:06:30.900339446 +0000 UTC m=+121.856981141" Dec 02 13:06:31 crc kubenswrapper[4725]: I1202 13:06:31.013649 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vs5fz"] Dec 02 13:06:31 crc kubenswrapper[4725]: I1202 13:06:31.013765 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:31 crc kubenswrapper[4725]: E1202 13:06:31.013846 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:31 crc kubenswrapper[4725]: I1202 13:06:31.267964 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:31 crc kubenswrapper[4725]: I1202 13:06:31.268057 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:31 crc kubenswrapper[4725]: E1202 13:06:31.268078 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:31 crc kubenswrapper[4725]: E1202 13:06:31.268194 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:32 crc kubenswrapper[4725]: I1202 13:06:32.267765 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:32 crc kubenswrapper[4725]: E1202 13:06:32.267894 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:33 crc kubenswrapper[4725]: I1202 13:06:33.267864 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:33 crc kubenswrapper[4725]: I1202 13:06:33.267934 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:33 crc kubenswrapper[4725]: I1202 13:06:33.267951 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:33 crc kubenswrapper[4725]: E1202 13:06:33.268022 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:33 crc kubenswrapper[4725]: E1202 13:06:33.268194 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:33 crc kubenswrapper[4725]: E1202 13:06:33.268240 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:34 crc kubenswrapper[4725]: I1202 13:06:34.268072 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:34 crc kubenswrapper[4725]: E1202 13:06:34.268263 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:34 crc kubenswrapper[4725]: I1202 13:06:34.268386 4725 scope.go:117] "RemoveContainer" containerID="9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913" Dec 02 13:06:34 crc kubenswrapper[4725]: E1202 13:06:34.392188 4725 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 13:06:34 crc kubenswrapper[4725]: I1202 13:06:34.886399 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/1.log" Dec 02 13:06:34 crc kubenswrapper[4725]: I1202 13:06:34.886442 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerStarted","Data":"d654f554b44b0403cae91f5d8143c150d9428226b7a282a7666f50eab5f10622"} Dec 02 13:06:35 crc kubenswrapper[4725]: I1202 13:06:35.267189 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:35 crc kubenswrapper[4725]: I1202 13:06:35.267300 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:35 crc kubenswrapper[4725]: E1202 13:06:35.267330 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:35 crc kubenswrapper[4725]: I1202 13:06:35.267189 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:35 crc kubenswrapper[4725]: E1202 13:06:35.267500 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:35 crc kubenswrapper[4725]: E1202 13:06:35.267602 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:36 crc kubenswrapper[4725]: I1202 13:06:36.268001 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:36 crc kubenswrapper[4725]: E1202 13:06:36.268122 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:37 crc kubenswrapper[4725]: I1202 13:06:37.268239 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:37 crc kubenswrapper[4725]: I1202 13:06:37.268406 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:37 crc kubenswrapper[4725]: E1202 13:06:37.268538 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:37 crc kubenswrapper[4725]: I1202 13:06:37.268690 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:37 crc kubenswrapper[4725]: E1202 13:06:37.268739 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:37 crc kubenswrapper[4725]: E1202 13:06:37.268911 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:38 crc kubenswrapper[4725]: I1202 13:06:38.267158 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:38 crc kubenswrapper[4725]: E1202 13:06:38.267296 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 13:06:39 crc kubenswrapper[4725]: I1202 13:06:39.268067 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:39 crc kubenswrapper[4725]: I1202 13:06:39.268080 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:39 crc kubenswrapper[4725]: I1202 13:06:39.268163 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:39 crc kubenswrapper[4725]: E1202 13:06:39.268894 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vs5fz" podUID="9feadd7b-f199-41c3-b908-43774e1cd2f3" Dec 02 13:06:39 crc kubenswrapper[4725]: E1202 13:06:39.268977 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 13:06:39 crc kubenswrapper[4725]: E1202 13:06:39.269091 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 13:06:40 crc kubenswrapper[4725]: I1202 13:06:40.267553 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:40 crc kubenswrapper[4725]: I1202 13:06:40.272067 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 02 13:06:40 crc kubenswrapper[4725]: I1202 13:06:40.272489 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.267598 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.267672 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.267782 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.271272 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.271309 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.271396 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 02 13:06:41 crc kubenswrapper[4725]: I1202 13:06:41.272728 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.853016 4725 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.884952 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h8fhj"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.885517 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.888508 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pb5sr"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.889009 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.894074 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.894236 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.900183 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.900352 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.909283 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.909884 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.910753 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.910912 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.913332 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.913551 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.913741 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.915741 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wst96"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.916224 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.917868 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.921345 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.921613 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.921815 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.922124 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.922325 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q9s7m"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.922531 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.922573 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.924085 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.924315 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.924422 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.927436 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.927882 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-gp8pw"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.929127 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.929489 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.929538 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.929805 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.929992 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.931040 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.934194 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.934124 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.934767 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.934856 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.935001 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.934951 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.935163 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.935174 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.935406 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.935736 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.936244 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7z7tx"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.936301 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.936687 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.936713 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.941209 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.943203 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-76gpw"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.943980 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.944235 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.944703 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.945566 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.945686 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.946700 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.948954 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.949139 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.959328 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hlts9"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.961521 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.979135 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.979627 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980085 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c17fbfab-7855-44b3-a983-22915062793a-config\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980134 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b286cd9d-5c21-43f8-a37d-469a65ec0815-serving-cert\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980170 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx66n\" (UniqueName: \"kubernetes.io/projected/b286cd9d-5c21-43f8-a37d-469a65ec0815-kube-api-access-hx66n\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980246 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b286cd9d-5c21-43f8-a37d-469a65ec0815-config\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980271 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c17fbfab-7855-44b3-a983-22915062793a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980291 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b286cd9d-5c21-43f8-a37d-469a65ec0815-trusted-ca\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980319 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c17fbfab-7855-44b3-a983-22915062793a-images\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.980339 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chzrs\" (UniqueName: \"kubernetes.io/projected/c17fbfab-7855-44b3-a983-22915062793a-kube-api-access-chzrs\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.981014 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.981290 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.981706 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.982093 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.982750 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6fvsl"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.983241 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.983330 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.983613 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.983967 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.984938 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.985364 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.985449 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.985643 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.985750 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.985918 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.985490 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.986020 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.986164 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.988881 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.989357 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.994980 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995067 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995188 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5m8fq"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995317 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995432 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995445 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995593 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995668 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995692 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995747 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.995768 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996011 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996092 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996316 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996378 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996427 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996507 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996545 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996607 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996619 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7wgpl"] Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996675 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996760 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996816 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996868 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.996916 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 02 13:06:44 crc kubenswrapper[4725]: I1202 13:06:44.997143 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.001742 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.001807 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002115 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002205 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002262 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002382 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002521 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002654 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.002943 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.003137 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.003217 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jx7zr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.003695 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004057 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gp8pw"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004089 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.003222 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004416 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004474 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004415 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004593 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004682 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004772 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.004974 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.005210 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.005452 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.005454 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.005894 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.007337 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.009444 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.010073 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.010697 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.012234 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.014034 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.014500 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.014682 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.016507 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.033283 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.034227 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.041108 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.041252 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.041953 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vngft"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.042321 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.042651 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.043895 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.044082 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.045255 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.045537 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fwlnc"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.047491 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.048589 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.049032 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q9s7m"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.050452 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.050970 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.051693 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.053056 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.053658 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.054081 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.054614 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.056599 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.056991 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.057136 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.057439 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.058931 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.059317 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.059356 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldhmm"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.059985 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.060820 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-knpzb"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.061572 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.062184 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.062592 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.063191 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.063756 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.065310 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zdq97"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.065689 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-llqpv"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.066077 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.066239 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.066331 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.068950 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pb5sr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.068981 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hlts9"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.069335 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.070348 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h8fhj"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.070675 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6fvsl"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.071662 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7z7tx"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.072793 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7wgpl"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.073736 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jx7zr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.074730 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.075695 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.076608 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-4jfjh"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.077499 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.077772 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.078902 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-76gpw"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.079925 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5m8fq"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.080963 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-config\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.080992 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-serving-cert\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081013 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b286cd9d-5c21-43f8-a37d-469a65ec0815-config\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081034 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4a1e44b-57f8-49bd-b67f-468593556909-config\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081050 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-etcd-client\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081083 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081131 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-serving-cert\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081280 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46x5c\" (UniqueName: \"kubernetes.io/projected/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-kube-api-access-46x5c\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081355 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln5wg\" (UniqueName: \"kubernetes.io/projected/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-kube-api-access-ln5wg\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081400 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c17fbfab-7855-44b3-a983-22915062793a-images\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081419 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx9n8\" (UniqueName: \"kubernetes.io/projected/a49031e7-e7da-47c1-b069-a5f129e2794f-kube-api-access-jx9n8\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081443 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a1e44b-57f8-49bd-b67f-468593556909-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081491 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081512 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnn4v\" (UniqueName: \"kubernetes.io/projected/bb2e34c8-7e96-44c9-a073-87101442f86e-kube-api-access-tnn4v\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081528 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-config\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081548 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b286cd9d-5c21-43f8-a37d-469a65ec0815-serving-cert\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081565 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-encryption-config\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081595 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx66n\" (UniqueName: \"kubernetes.io/projected/b286cd9d-5c21-43f8-a37d-469a65ec0815-kube-api-access-hx66n\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081610 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-etcd-serving-ca\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081624 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb2e34c8-7e96-44c9-a073-87101442f86e-audit-dir\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081654 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-config\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081681 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081699 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c17fbfab-7855-44b3-a983-22915062793a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081723 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b286cd9d-5c21-43f8-a37d-469a65ec0815-trusted-ca\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081740 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f4a1e44b-57f8-49bd-b67f-468593556909-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081756 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c89qx\" (UID: \"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081774 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49031e7-e7da-47c1-b069-a5f129e2794f-serving-cert\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081788 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-image-import-ca\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081806 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-service-ca-bundle\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081822 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081839 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chzrs\" (UniqueName: \"kubernetes.io/projected/c17fbfab-7855-44b3-a983-22915062793a-kube-api-access-chzrs\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081855 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c17fbfab-7855-44b3-a983-22915062793a-config\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081873 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6sd7\" (UniqueName: \"kubernetes.io/projected/9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936-kube-api-access-p6sd7\") pod \"cluster-samples-operator-665b6dd947-c89qx\" (UID: \"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081888 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-audit\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081904 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b286cd9d-5c21-43f8-a37d-469a65ec0815-config\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081911 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-serving-cert\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.081985 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.082010 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bb2e34c8-7e96-44c9-a073-87101442f86e-node-pullsecrets\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.082025 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-client-ca\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.082171 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c17fbfab-7855-44b3-a983-22915062793a-images\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.082851 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c17fbfab-7855-44b3-a983-22915062793a-config\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.083067 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b286cd9d-5c21-43f8-a37d-469a65ec0815-trusted-ca\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.085360 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.086326 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.087510 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.091057 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.091110 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vngft"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.091121 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.095407 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fwlnc"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.095441 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wst96"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.095451 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.099758 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.099807 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.099821 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.102995 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.103187 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.103222 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-knpzb"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.103531 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c17fbfab-7855-44b3-a983-22915062793a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.105243 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldhmm"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.106492 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.108224 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b286cd9d-5c21-43f8-a37d-469a65ec0815-serving-cert\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.108293 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.111523 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.116741 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2kwpr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.120610 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4jfjh"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.120648 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-llqpv"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.120660 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.120755 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.130255 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2kwpr"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.143493 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8wm9d"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.144220 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.144974 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g49dl"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.146273 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.146330 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g49dl"] Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.147692 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.169019 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182637 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-etcd-client\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182679 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46x5c\" (UniqueName: \"kubernetes.io/projected/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-kube-api-access-46x5c\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182747 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-serving-cert\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182775 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln5wg\" (UniqueName: \"kubernetes.io/projected/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-kube-api-access-ln5wg\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182812 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx9n8\" (UniqueName: \"kubernetes.io/projected/a49031e7-e7da-47c1-b069-a5f129e2794f-kube-api-access-jx9n8\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182836 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a1e44b-57f8-49bd-b67f-468593556909-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182857 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182881 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnn4v\" (UniqueName: \"kubernetes.io/projected/bb2e34c8-7e96-44c9-a073-87101442f86e-kube-api-access-tnn4v\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182902 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-config\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182923 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-encryption-config\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182952 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-etcd-serving-ca\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.182972 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb2e34c8-7e96-44c9-a073-87101442f86e-audit-dir\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183029 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-config\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183053 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183074 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f4a1e44b-57f8-49bd-b67f-468593556909-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183097 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c89qx\" (UID: \"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183118 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-image-import-ca\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183151 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49031e7-e7da-47c1-b069-a5f129e2794f-serving-cert\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183174 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-service-ca-bundle\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183200 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183232 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6sd7\" (UniqueName: \"kubernetes.io/projected/9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936-kube-api-access-p6sd7\") pod \"cluster-samples-operator-665b6dd947-c89qx\" (UID: \"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183250 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-audit\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183272 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-serving-cert\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183302 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183321 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bb2e34c8-7e96-44c9-a073-87101442f86e-node-pullsecrets\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183336 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-client-ca\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183352 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-config\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183368 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-serving-cert\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183384 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4a1e44b-57f8-49bd-b67f-468593556909-config\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183414 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.183984 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-config\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.184081 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-service-ca-bundle\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.184380 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-config\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.184529 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a49031e7-e7da-47c1-b069-a5f129e2794f-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.184590 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-client-ca\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.184649 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/bb2e34c8-7e96-44c9-a073-87101442f86e-node-pullsecrets\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.184677 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb2e34c8-7e96-44c9-a073-87101442f86e-audit-dir\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.185285 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-etcd-serving-ca\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.185500 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.185609 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-image-import-ca\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.186151 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-etcd-client\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.186522 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.186854 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-config\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.187064 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c89qx\" (UID: \"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.188212 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.189293 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a49031e7-e7da-47c1-b069-a5f129e2794f-serving-cert\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.189868 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-serving-cert\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.195787 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-serving-cert\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.196187 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-encryption-config\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.196231 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb2e34c8-7e96-44c9-a073-87101442f86e-serving-cert\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.196266 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/bb2e34c8-7e96-44c9-a073-87101442f86e-audit\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.207286 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.228408 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.247806 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.267953 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.287480 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.307794 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.328246 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.347787 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.367530 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.388332 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.407286 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.416141 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f4a1e44b-57f8-49bd-b67f-468593556909-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.427553 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.434451 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4a1e44b-57f8-49bd-b67f-468593556909-config\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.468182 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.488400 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.507889 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.534075 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.547885 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.567978 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.588823 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.607986 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.627637 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.648437 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.668518 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.687758 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.707911 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.728362 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.748906 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.768430 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.788273 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.808237 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.828353 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.847865 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.867790 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.887758 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.908206 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.928349 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.948233 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.967370 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 02 13:06:45 crc kubenswrapper[4725]: I1202 13:06:45.988041 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.008064 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.028348 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.047822 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.066236 4725 request.go:700] Waited for 1.01128477s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.068070 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.087398 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.108283 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.128037 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.148275 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.168507 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.187651 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.214853 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.228792 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.247485 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.268395 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.288439 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.308833 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.328038 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.348719 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.368351 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.388203 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.407195 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.428340 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.447863 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.467987 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.488099 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.507666 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.527903 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.547912 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.568381 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.588254 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.607249 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.627211 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.647826 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.667896 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.687120 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.708835 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.728556 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.747450 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.786524 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx66n\" (UniqueName: \"kubernetes.io/projected/b286cd9d-5c21-43f8-a37d-469a65ec0815-kube-api-access-hx66n\") pod \"console-operator-58897d9998-pb5sr\" (UID: \"b286cd9d-5c21-43f8-a37d-469a65ec0815\") " pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.803721 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chzrs\" (UniqueName: \"kubernetes.io/projected/c17fbfab-7855-44b3-a983-22915062793a-kube-api-access-chzrs\") pod \"machine-api-operator-5694c8668f-h8fhj\" (UID: \"c17fbfab-7855-44b3-a983-22915062793a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.808040 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.828344 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.847261 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.867866 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.907621 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.928366 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.948190 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.968105 4725 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 02 13:06:46 crc kubenswrapper[4725]: I1202 13:06:46.987702 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.007318 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.027697 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.046017 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.049592 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46x5c\" (UniqueName: \"kubernetes.io/projected/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-kube-api-access-46x5c\") pod \"controller-manager-879f6c89f-q9s7m\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.062351 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx9n8\" (UniqueName: \"kubernetes.io/projected/a49031e7-e7da-47c1-b069-a5f129e2794f-kube-api-access-jx9n8\") pod \"authentication-operator-69f744f599-hlts9\" (UID: \"a49031e7-e7da-47c1-b069-a5f129e2794f\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.066827 4725 request.go:700] Waited for 1.883806637s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-config-operator/serviceaccounts/openshift-config-operator/token Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.082597 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln5wg\" (UniqueName: \"kubernetes.io/projected/87c41c9b-efe7-40ba-867c-1fd805f5ff7e-kube-api-access-ln5wg\") pod \"openshift-config-operator-7777fb866f-76gpw\" (UID: \"87c41c9b-efe7-40ba-867c-1fd805f5ff7e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.102599 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnn4v\" (UniqueName: \"kubernetes.io/projected/bb2e34c8-7e96-44c9-a073-87101442f86e-kube-api-access-tnn4v\") pod \"apiserver-76f77b778f-6fvsl\" (UID: \"bb2e34c8-7e96-44c9-a073-87101442f86e\") " pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.127747 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f4a1e44b-57f8-49bd-b67f-468593556909-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-hgcrs\" (UID: \"f4a1e44b-57f8-49bd-b67f-468593556909\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.142526 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.145449 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6sd7\" (UniqueName: \"kubernetes.io/projected/9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936-kube-api-access-p6sd7\") pod \"cluster-samples-operator-665b6dd947-c89qx\" (UID: \"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.166272 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.169928 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.184077 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203413 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-trusted-ca\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203487 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203517 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghk2k\" (UniqueName: \"kubernetes.io/projected/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-kube-api-access-ghk2k\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203572 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6g8l\" (UniqueName: \"kubernetes.io/projected/48f14ad4-c071-4e02-8182-82ffd354c74f-kube-api-access-v6g8l\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203657 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-client\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203681 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2b66987-71e9-472b-9d08-5e8d89dfe129-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203707 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28fp\" (UniqueName: \"kubernetes.io/projected/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-kube-api-access-g28fp\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203761 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f14ad4-c071-4e02-8182-82ffd354c74f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203783 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-encryption-config\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203848 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203899 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-serving-cert\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.203931 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v98kc\" (UniqueName: \"kubernetes.io/projected/e2b66987-71e9-472b-9d08-5e8d89dfe129-kube-api-access-v98kc\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204063 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7828c47b-d717-48f4-ba0a-375358ec3d79-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204091 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6-metrics-tls\") pod \"dns-operator-744455d44c-jx7zr\" (UID: \"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204115 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-service-ca\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204166 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75642\" (UniqueName: \"kubernetes.io/projected/f3877ac3-8496-4b4f-9578-8ee270d55db2-kube-api-access-75642\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204225 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f3877ac3-8496-4b4f-9578-8ee270d55db2-audit-dir\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.204229 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:47.7042137 +0000 UTC m=+138.660855505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204261 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2b66987-71e9-472b-9d08-5e8d89dfe129-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204309 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204310 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79psm\" (UniqueName: \"kubernetes.io/projected/e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6-kube-api-access-79psm\") pod \"dns-operator-744455d44c-jx7zr\" (UID: \"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204520 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-console-config\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204546 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlhfd\" (UniqueName: \"kubernetes.io/projected/2034b4c1-5d0f-42d4-951d-afd68a259ee1-kube-api-access-xlhfd\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204571 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1ed82a59-61d2-430c-b819-734c1c2f00b1-metrics-tls\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204594 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-service-ca\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204616 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-serving-cert\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204638 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-etcd-client\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204661 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ed82a59-61d2-430c-b819-734c1c2f00b1-trusted-ca\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204705 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-bound-sa-token\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204729 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204764 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204912 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49vjn\" (UniqueName: \"kubernetes.io/projected/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-kube-api-access-49vjn\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204956 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-oauth-serving-cert\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.204990 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-client-ca\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205016 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-config\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205041 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-config\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205065 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-serving-cert\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205087 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-oauth-config\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205137 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-ca\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205164 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-audit-policies\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205261 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-tls\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205548 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205575 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzdzn\" (UniqueName: \"kubernetes.io/projected/1ed82a59-61d2-430c-b819-734c1c2f00b1-kube-api-access-rzdzn\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205618 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-machine-approver-tls\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205641 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpxwp\" (UniqueName: \"kubernetes.io/projected/a81f9c3a-432a-4454-95fa-aea3a8cc3344-kube-api-access-wpxwp\") pod \"downloads-7954f5f757-gp8pw\" (UID: \"a81f9c3a-432a-4454-95fa-aea3a8cc3344\") " pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205698 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmrpd\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-kube-api-access-fmrpd\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205916 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24dca536-7872-4a56-91cc-3361f59b57cf-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205946 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24dca536-7872-4a56-91cc-3361f59b57cf-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205971 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.205995 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-config\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206027 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206064 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-dir\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206089 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206128 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206186 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206234 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-policies\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206262 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206287 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206311 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-serving-cert\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206331 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ed82a59-61d2-430c-b819-734c1c2f00b1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206355 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f14ad4-c071-4e02-8182-82ffd354c74f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206397 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-certificates\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206418 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/24dca536-7872-4a56-91cc-3361f59b57cf-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206439 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbxw\" (UniqueName: \"kubernetes.io/projected/24dca536-7872-4a56-91cc-3361f59b57cf-kube-api-access-vxbxw\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206477 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lzdh\" (UniqueName: \"kubernetes.io/projected/7971baa9-a25e-47e2-ac21-8c89db535872-kube-api-access-2lzdh\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206532 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7828c47b-d717-48f4-ba0a-375358ec3d79-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206557 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-auth-proxy-config\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206573 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-trusted-ca-bundle\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206594 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.206620 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.280390 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-h8fhj"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.287720 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pb5sr"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.307889 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308100 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f3877ac3-8496-4b4f-9578-8ee270d55db2-audit-dir\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308119 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2b66987-71e9-472b-9d08-5e8d89dfe129-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308142 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-proxy-tls\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308157 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a128839-ee0f-4a13-b04d-422d4c76f9ee-service-ca-bundle\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308173 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-etcd-client\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308187 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-service-ca\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308206 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-serving-cert\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308248 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308265 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6db27d13-bcf4-4596-96ba-5f140c9fb38b-srv-cert\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308289 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49vjn\" (UniqueName: \"kubernetes.io/projected/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-kube-api-access-49vjn\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308303 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-oauth-serving-cert\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308319 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-webhook-cert\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308334 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18606d46-7784-43be-826b-1b1266255201-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fwlnc\" (UID: \"18606d46-7784-43be-826b-1b1266255201\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308349 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dscpw\" (UniqueName: \"kubernetes.io/projected/4bb9a525-a393-4552-83f9-22ceca686349-kube-api-access-dscpw\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308365 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-config\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308379 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-config\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308396 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-serving-cert\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308410 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-oauth-config\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308426 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk4gk\" (UniqueName: \"kubernetes.io/projected/bd10ea21-eff4-4fe0-9474-02d16b2d135d-kube-api-access-kk4gk\") pod \"ingress-canary-2kwpr\" (UID: \"bd10ea21-eff4-4fe0-9474-02d16b2d135d\") " pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308441 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-csi-data-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308504 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-audit-policies\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308524 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62013aee-5ac1-47d6-a5fa-cdadd084ebad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308538 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-tls\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308555 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308572 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-machine-approver-tls\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308594 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-stats-auth\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308611 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308630 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308644 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-plugins-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308658 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-serving-cert\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308673 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4bb9a525-a393-4552-83f9-22ceca686349-signing-cabundle\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308690 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-dir\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308706 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308721 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-socket-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308739 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc2dl\" (UniqueName: \"kubernetes.io/projected/7900355d-cbde-4a76-91ef-9339769983f9-kube-api-access-pc2dl\") pod \"package-server-manager-789f6589d5-pg7x8\" (UID: \"7900355d-cbde-4a76-91ef-9339769983f9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308755 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-config\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308777 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308794 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-policies\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308810 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308829 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-serving-cert\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308845 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxtxl\" (UniqueName: \"kubernetes.io/projected/66462d0f-425e-470b-a4e2-94682d534fe9-kube-api-access-hxtxl\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308860 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lzdh\" (UniqueName: \"kubernetes.io/projected/7971baa9-a25e-47e2-ac21-8c89db535872-kube-api-access-2lzdh\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308875 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx286\" (UniqueName: \"kubernetes.io/projected/6db27d13-bcf4-4596-96ba-5f140c9fb38b-kube-api-access-wx286\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308890 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7828c47b-d717-48f4-ba0a-375358ec3d79-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308904 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-auth-proxy-config\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308921 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/04180a2f-4341-498d-9bd0-485967ae26cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308936 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/233aca62-4621-42ea-b50b-ee87e52af8a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308968 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.308991 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309006 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-trusted-ca\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309021 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309045 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-metrics-certs\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309062 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slg75\" (UniqueName: \"kubernetes.io/projected/68f88274-d112-4899-bc9d-e9766a6478d3-kube-api-access-slg75\") pod \"migrator-59844c95c7-hn6hw\" (UID: \"68f88274-d112-4899-bc9d-e9766a6478d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309076 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04180a2f-4341-498d-9bd0-485967ae26cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309091 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6db27d13-bcf4-4596-96ba-5f140c9fb38b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309107 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2b66987-71e9-472b-9d08-5e8d89dfe129-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309122 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl82h\" (UniqueName: \"kubernetes.io/projected/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-kube-api-access-hl82h\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309147 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f14ad4-c071-4e02-8182-82ffd354c74f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309162 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-encryption-config\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309177 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04180a2f-4341-498d-9bd0-485967ae26cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309193 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-serving-cert\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309208 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v98kc\" (UniqueName: \"kubernetes.io/projected/e2b66987-71e9-472b-9d08-5e8d89dfe129-kube-api-access-v98kc\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309240 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ea95c612-3884-488c-a153-9c6d32cc54aa-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309256 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-service-ca\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309270 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75642\" (UniqueName: \"kubernetes.io/projected/f3877ac3-8496-4b4f-9578-8ee270d55db2-kube-api-access-75642\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309296 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzjv4\" (UniqueName: \"kubernetes.io/projected/233aca62-4621-42ea-b50b-ee87e52af8a9-kube-api-access-qzjv4\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309311 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6d2x\" (UniqueName: \"kubernetes.io/projected/44387809-9d4a-43c6-b61d-364ab1b6c48f-kube-api-access-x6d2x\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309336 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79psm\" (UniqueName: \"kubernetes.io/projected/e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6-kube-api-access-79psm\") pod \"dns-operator-744455d44c-jx7zr\" (UID: \"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309351 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-console-config\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309367 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjn4c\" (UniqueName: \"kubernetes.io/projected/18606d46-7784-43be-826b-1b1266255201-kube-api-access-kjn4c\") pod \"multus-admission-controller-857f4d67dd-fwlnc\" (UID: \"18606d46-7784-43be-826b-1b1266255201\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309392 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlhfd\" (UniqueName: \"kubernetes.io/projected/2034b4c1-5d0f-42d4-951d-afd68a259ee1-kube-api-access-xlhfd\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309418 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79n7q\" (UniqueName: \"kubernetes.io/projected/ea95c612-3884-488c-a153-9c6d32cc54aa-kube-api-access-79n7q\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309435 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ngk2\" (UniqueName: \"kubernetes.io/projected/2abf7968-024b-4d46-882a-0d8029b116e2-kube-api-access-6ngk2\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309450 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b0990e-4363-4b28-88fa-dfde258940ac-config-volume\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309487 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1ed82a59-61d2-430c-b819-734c1c2f00b1-metrics-tls\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309503 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ed82a59-61d2-430c-b819-734c1c2f00b1-trusted-ca\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309518 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-bound-sa-token\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.309580 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:47.809548903 +0000 UTC m=+138.766190608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309638 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309711 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2btc\" (UniqueName: \"kubernetes.io/projected/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-kube-api-access-w2btc\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309741 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2abf7968-024b-4d46-882a-0d8029b116e2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309785 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66462d0f-425e-470b-a4e2-94682d534fe9-config-volume\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309816 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-client-ca\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309840 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-registration-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309865 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ea95c612-3884-488c-a153-9c6d32cc54aa-proxy-tls\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309889 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66462d0f-425e-470b-a4e2-94682d534fe9-secret-volume\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309913 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4bb9a525-a393-4552-83f9-22ceca686349-signing-key\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309943 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-ca\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309969 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpxwp\" (UniqueName: \"kubernetes.io/projected/a81f9c3a-432a-4454-95fa-aea3a8cc3344-kube-api-access-wpxwp\") pod \"downloads-7954f5f757-gp8pw\" (UID: \"a81f9c3a-432a-4454-95fa-aea3a8cc3344\") " pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.309993 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/77e43a6d-f759-4340-a285-46c85b8044cd-certs\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310048 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzdzn\" (UniqueName: \"kubernetes.io/projected/1ed82a59-61d2-430c-b819-734c1c2f00b1-kube-api-access-rzdzn\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310078 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrpd\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-kube-api-access-fmrpd\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310104 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24dca536-7872-4a56-91cc-3361f59b57cf-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310130 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24dca536-7872-4a56-91cc-3361f59b57cf-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310156 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310201 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-config\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310224 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd10ea21-eff4-4fe0-9474-02d16b2d135d-cert\") pod \"ingress-canary-2kwpr\" (UID: \"bd10ea21-eff4-4fe0-9474-02d16b2d135d\") " pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310245 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2x5f\" (UniqueName: \"kubernetes.io/projected/77e43a6d-f759-4340-a285-46c85b8044cd-kube-api-access-q2x5f\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310285 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkdb4\" (UniqueName: \"kubernetes.io/projected/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-kube-api-access-kkdb4\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310324 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310350 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310379 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ed82a59-61d2-430c-b819-734c1c2f00b1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310409 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f14ad4-c071-4e02-8182-82ffd354c74f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310435 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/233aca62-4621-42ea-b50b-ee87e52af8a9-srv-cert\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310494 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/24dca536-7872-4a56-91cc-3361f59b57cf-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310521 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbxw\" (UniqueName: \"kubernetes.io/projected/24dca536-7872-4a56-91cc-3361f59b57cf-kube-api-access-vxbxw\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310548 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fg6t\" (UniqueName: \"kubernetes.io/projected/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-kube-api-access-5fg6t\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310574 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-certificates\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310598 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-trusted-ca-bundle\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310622 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-apiservice-cert\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310655 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62013aee-5ac1-47d6-a5fa-cdadd084ebad-config\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310697 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6g8l\" (UniqueName: \"kubernetes.io/projected/48f14ad4-c071-4e02-8182-82ffd354c74f-kube-api-access-v6g8l\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310726 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310749 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62013aee-5ac1-47d6-a5fa-cdadd084ebad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310777 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310801 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghk2k\" (UniqueName: \"kubernetes.io/projected/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-kube-api-access-ghk2k\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310827 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzqsz\" (UniqueName: \"kubernetes.io/projected/96b0990e-4363-4b28-88fa-dfde258940ac-kube-api-access-rzqsz\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310828 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-policies\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310869 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d37326e-fdd1-41e0-8401-5f47038356b3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bphpt\" (UID: \"4d37326e-fdd1-41e0-8401-5f47038356b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310898 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ea95c612-3884-488c-a153-9c6d32cc54aa-images\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310921 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf4s5\" (UniqueName: \"kubernetes.io/projected/7a128839-ee0f-4a13-b04d-422d4c76f9ee-kube-api-access-bf4s5\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310949 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-client\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310972 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7900355d-cbde-4a76-91ef-9339769983f9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pg7x8\" (UID: \"7900355d-cbde-4a76-91ef-9339769983f9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.310998 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-default-certificate\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.311024 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28fp\" (UniqueName: \"kubernetes.io/projected/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-kube-api-access-g28fp\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.311048 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/77e43a6d-f759-4340-a285-46c85b8044cd-node-bootstrap-token\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.311084 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-mountpoint-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.311106 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96c99\" (UniqueName: \"kubernetes.io/projected/4d37326e-fdd1-41e0-8401-5f47038356b3-kube-api-access-96c99\") pod \"control-plane-machine-set-operator-78cbb6b69f-bphpt\" (UID: \"4d37326e-fdd1-41e0-8401-5f47038356b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.311173 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abf7968-024b-4d46-882a-0d8029b116e2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.311900 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.312350 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.312507 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f3877ac3-8496-4b4f-9578-8ee270d55db2-audit-dir\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.313053 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2b66987-71e9-472b-9d08-5e8d89dfe129-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.313377 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-console-config\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.313543 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.313577 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-tmpfs\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.313599 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-config\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.314210 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.314477 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-auth-proxy-config\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.322051 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-config\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.322877 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-encryption-config\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.322925 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.326061 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f14ad4-c071-4e02-8182-82ffd354c74f-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.327019 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-serving-cert\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.327689 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-ca\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.328278 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.334227 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-audit-policies\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.334504 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-serving-cert\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.334929 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7828c47b-d717-48f4-ba0a-375358ec3d79-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.335894 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ed82a59-61d2-430c-b819-734c1c2f00b1-trusted-ca\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.336122 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3877ac3-8496-4b4f-9578-8ee270d55db2-etcd-client\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.336645 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-service-ca\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.337836 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24dca536-7872-4a56-91cc-3361f59b57cf-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.339506 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1ed82a59-61d2-430c-b819-734c1c2f00b1-metrics-tls\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.339627 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-config\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.339925 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-dir\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.341832 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3877ac3-8496-4b4f-9578-8ee270d55db2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.343123 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-oauth-config\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.345572 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-client-ca\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.347052 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-oauth-serving-cert\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.347948 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-service-ca\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.348596 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.349277 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.351158 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.352253 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f14ad4-c071-4e02-8182-82ffd354c74f-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.353695 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-serving-cert\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.354642 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-serving-cert\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.354981 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.355296 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-machine-approver-tls\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.355544 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.355798 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.358020 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-trusted-ca-bundle\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.358700 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-bound-sa-token\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.358706 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-certificates\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.358791 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b0990e-4363-4b28-88fa-dfde258940ac-metrics-tls\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.358797 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-etcd-client\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.359063 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7828c47b-d717-48f4-ba0a-375358ec3d79-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.359145 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-tls\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.359501 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2b66987-71e9-472b-9d08-5e8d89dfe129-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.359553 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:47.859527724 +0000 UTC m=+138.816169419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.360536 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.363440 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-trusted-ca\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.363911 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/24dca536-7872-4a56-91cc-3361f59b57cf-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.365282 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6-metrics-tls\") pod \"dns-operator-744455d44c-jx7zr\" (UID: \"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.366388 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7828c47b-d717-48f4-ba0a-375358ec3d79-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.368682 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.368743 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6-metrics-tls\") pod \"dns-operator-744455d44c-jx7zr\" (UID: \"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.382686 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79psm\" (UniqueName: \"kubernetes.io/projected/e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6-kube-api-access-79psm\") pod \"dns-operator-744455d44c-jx7zr\" (UID: \"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6\") " pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.403593 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlhfd\" (UniqueName: \"kubernetes.io/projected/2034b4c1-5d0f-42d4-951d-afd68a259ee1-kube-api-access-xlhfd\") pod \"oauth-openshift-558db77b4-wst96\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.425205 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lzdh\" (UniqueName: \"kubernetes.io/projected/7971baa9-a25e-47e2-ac21-8c89db535872-kube-api-access-2lzdh\") pod \"console-f9d7485db-7z7tx\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.444880 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v98kc\" (UniqueName: \"kubernetes.io/projected/e2b66987-71e9-472b-9d08-5e8d89dfe129-kube-api-access-v98kc\") pod \"openshift-controller-manager-operator-756b6f6bc6-f4r68\" (UID: \"e2b66987-71e9-472b-9d08-5e8d89dfe129\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.464422 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpxwp\" (UniqueName: \"kubernetes.io/projected/a81f9c3a-432a-4454-95fa-aea3a8cc3344-kube-api-access-wpxwp\") pod \"downloads-7954f5f757-gp8pw\" (UID: \"a81f9c3a-432a-4454-95fa-aea3a8cc3344\") " pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466134 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.466549 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:47.966521037 +0000 UTC m=+138.923162732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466593 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-metrics-certs\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466628 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slg75\" (UniqueName: \"kubernetes.io/projected/68f88274-d112-4899-bc9d-e9766a6478d3-kube-api-access-slg75\") pod \"migrator-59844c95c7-hn6hw\" (UID: \"68f88274-d112-4899-bc9d-e9766a6478d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466649 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04180a2f-4341-498d-9bd0-485967ae26cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466665 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6db27d13-bcf4-4596-96ba-5f140c9fb38b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466685 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl82h\" (UniqueName: \"kubernetes.io/projected/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-kube-api-access-hl82h\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466709 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04180a2f-4341-498d-9bd0-485967ae26cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466735 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ea95c612-3884-488c-a153-9c6d32cc54aa-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466761 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzjv4\" (UniqueName: \"kubernetes.io/projected/233aca62-4621-42ea-b50b-ee87e52af8a9-kube-api-access-qzjv4\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466815 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6d2x\" (UniqueName: \"kubernetes.io/projected/44387809-9d4a-43c6-b61d-364ab1b6c48f-kube-api-access-x6d2x\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466837 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjn4c\" (UniqueName: \"kubernetes.io/projected/18606d46-7784-43be-826b-1b1266255201-kube-api-access-kjn4c\") pod \"multus-admission-controller-857f4d67dd-fwlnc\" (UID: \"18606d46-7784-43be-826b-1b1266255201\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466867 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79n7q\" (UniqueName: \"kubernetes.io/projected/ea95c612-3884-488c-a153-9c6d32cc54aa-kube-api-access-79n7q\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466883 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ngk2\" (UniqueName: \"kubernetes.io/projected/2abf7968-024b-4d46-882a-0d8029b116e2-kube-api-access-6ngk2\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466901 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b0990e-4363-4b28-88fa-dfde258940ac-config-volume\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466923 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2btc\" (UniqueName: \"kubernetes.io/projected/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-kube-api-access-w2btc\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466938 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2abf7968-024b-4d46-882a-0d8029b116e2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466966 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66462d0f-425e-470b-a4e2-94682d534fe9-config-volume\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.466985 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-registration-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.467002 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ea95c612-3884-488c-a153-9c6d32cc54aa-proxy-tls\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.467017 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66462d0f-425e-470b-a4e2-94682d534fe9-secret-volume\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.467033 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4bb9a525-a393-4552-83f9-22ceca686349-signing-key\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.467062 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/77e43a6d-f759-4340-a285-46c85b8044cd-certs\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.467100 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd10ea21-eff4-4fe0-9474-02d16b2d135d-cert\") pod \"ingress-canary-2kwpr\" (UID: \"bd10ea21-eff4-4fe0-9474-02d16b2d135d\") " pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.467118 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2x5f\" (UniqueName: \"kubernetes.io/projected/77e43a6d-f759-4340-a285-46c85b8044cd-kube-api-access-q2x5f\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.468015 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04180a2f-4341-498d-9bd0-485967ae26cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.468233 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96b0990e-4363-4b28-88fa-dfde258940ac-config-volume\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.468298 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ea95c612-3884-488c-a153-9c6d32cc54aa-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.469130 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04180a2f-4341-498d-9bd0-485967ae26cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.469533 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-metrics-certs\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.469623 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-registration-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470110 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66462d0f-425e-470b-a4e2-94682d534fe9-config-volume\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470262 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkdb4\" (UniqueName: \"kubernetes.io/projected/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-kube-api-access-kkdb4\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470340 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/233aca62-4621-42ea-b50b-ee87e52af8a9-srv-cert\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470377 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fg6t\" (UniqueName: \"kubernetes.io/projected/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-kube-api-access-5fg6t\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470404 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-apiservice-cert\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470432 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62013aee-5ac1-47d6-a5fa-cdadd084ebad-config\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470512 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470541 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62013aee-5ac1-47d6-a5fa-cdadd084ebad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470568 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d37326e-fdd1-41e0-8401-5f47038356b3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bphpt\" (UID: \"4d37326e-fdd1-41e0-8401-5f47038356b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470593 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzqsz\" (UniqueName: \"kubernetes.io/projected/96b0990e-4363-4b28-88fa-dfde258940ac-kube-api-access-rzqsz\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470617 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf4s5\" (UniqueName: \"kubernetes.io/projected/7a128839-ee0f-4a13-b04d-422d4c76f9ee-kube-api-access-bf4s5\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470638 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ea95c612-3884-488c-a153-9c6d32cc54aa-images\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470660 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7900355d-cbde-4a76-91ef-9339769983f9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pg7x8\" (UID: \"7900355d-cbde-4a76-91ef-9339769983f9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470682 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/77e43a6d-f759-4340-a285-46c85b8044cd-node-bootstrap-token\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470702 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-default-certificate\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.470743 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-mountpoint-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.471349 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6db27d13-bcf4-4596-96ba-5f140c9fb38b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.471489 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66462d0f-425e-470b-a4e2-94682d534fe9-secret-volume\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.471623 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4bb9a525-a393-4552-83f9-22ceca686349-signing-key\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.471687 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62013aee-5ac1-47d6-a5fa-cdadd084ebad-config\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.471846 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ea95c612-3884-488c-a153-9c6d32cc54aa-images\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.472275 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ea95c612-3884-488c-a153-9c6d32cc54aa-proxy-tls\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473035 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96c99\" (UniqueName: \"kubernetes.io/projected/4d37326e-fdd1-41e0-8401-5f47038356b3-kube-api-access-96c99\") pod \"control-plane-machine-set-operator-78cbb6b69f-bphpt\" (UID: \"4d37326e-fdd1-41e0-8401-5f47038356b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473117 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-mountpoint-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473283 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abf7968-024b-4d46-882a-0d8029b116e2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473325 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-tmpfs\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473353 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473417 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b0990e-4363-4b28-88fa-dfde258940ac-metrics-tls\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473444 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a128839-ee0f-4a13-b04d-422d4c76f9ee-service-ca-bundle\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473491 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-proxy-tls\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473522 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6db27d13-bcf4-4596-96ba-5f140c9fb38b-srv-cert\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473554 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-webhook-cert\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473578 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18606d46-7784-43be-826b-1b1266255201-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fwlnc\" (UID: \"18606d46-7784-43be-826b-1b1266255201\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473625 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dscpw\" (UniqueName: \"kubernetes.io/projected/4bb9a525-a393-4552-83f9-22ceca686349-kube-api-access-dscpw\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473657 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk4gk\" (UniqueName: \"kubernetes.io/projected/bd10ea21-eff4-4fe0-9474-02d16b2d135d-kube-api-access-kk4gk\") pod \"ingress-canary-2kwpr\" (UID: \"bd10ea21-eff4-4fe0-9474-02d16b2d135d\") " pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473679 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-csi-data-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473708 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62013aee-5ac1-47d6-a5fa-cdadd084ebad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473743 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/233aca62-4621-42ea-b50b-ee87e52af8a9-srv-cert\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473749 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473797 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-stats-auth\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473819 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-plugins-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473837 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-serving-cert\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473854 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4bb9a525-a393-4552-83f9-22ceca686349-signing-cabundle\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473875 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-socket-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473895 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc2dl\" (UniqueName: \"kubernetes.io/projected/7900355d-cbde-4a76-91ef-9339769983f9-kube-api-access-pc2dl\") pod \"package-server-manager-789f6589d5-pg7x8\" (UID: \"7900355d-cbde-4a76-91ef-9339769983f9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473913 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-config\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473942 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxtxl\" (UniqueName: \"kubernetes.io/projected/66462d0f-425e-470b-a4e2-94682d534fe9-kube-api-access-hxtxl\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473959 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx286\" (UniqueName: \"kubernetes.io/projected/6db27d13-bcf4-4596-96ba-5f140c9fb38b-kube-api-access-wx286\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473980 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/04180a2f-4341-498d-9bd0-485967ae26cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.473995 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/233aca62-4621-42ea-b50b-ee87e52af8a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.474016 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.474075 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7900355d-cbde-4a76-91ef-9339769983f9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pg7x8\" (UID: \"7900355d-cbde-4a76-91ef-9339769983f9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.474323 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.474445 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.475130 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/77e43a6d-f759-4340-a285-46c85b8044cd-certs\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.475322 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.475595 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-apiservice-cert\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.475620 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.475863 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-config\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.476336 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4bb9a525-a393-4552-83f9-22ceca686349-signing-cabundle\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.476417 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-socket-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.477075 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2abf7968-024b-4d46-882a-0d8029b116e2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.477280 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:47.977266857 +0000 UTC m=+138.933908542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.477346 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-default-certificate\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.477420 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62013aee-5ac1-47d6-a5fa-cdadd084ebad-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.478080 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-serving-cert\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.478107 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-csi-data-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.478398 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-tmpfs\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.479022 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2abf7968-024b-4d46-882a-0d8029b116e2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.479594 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/233aca62-4621-42ea-b50b-ee87e52af8a9-profile-collector-cert\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.480025 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6db27d13-bcf4-4596-96ba-5f140c9fb38b-srv-cert\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.480170 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-plugins-dir\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.480271 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/77e43a6d-f759-4340-a285-46c85b8044cd-node-bootstrap-token\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.480859 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a128839-ee0f-4a13-b04d-422d4c76f9ee-service-ca-bundle\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.480891 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b0990e-4363-4b28-88fa-dfde258940ac-metrics-tls\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.481404 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-proxy-tls\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.481454 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a128839-ee0f-4a13-b04d-422d4c76f9ee-stats-auth\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.482218 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/18606d46-7784-43be-826b-1b1266255201-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fwlnc\" (UID: \"18606d46-7784-43be-826b-1b1266255201\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.482787 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4d37326e-fdd1-41e0-8401-5f47038356b3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bphpt\" (UID: \"4d37326e-fdd1-41e0-8401-5f47038356b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.484320 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bd10ea21-eff4-4fe0-9474-02d16b2d135d-cert\") pod \"ingress-canary-2kwpr\" (UID: \"bd10ea21-eff4-4fe0-9474-02d16b2d135d\") " pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.484895 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-webhook-cert\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.502050 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmrpd\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-kube-api-access-fmrpd\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.522282 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24dca536-7872-4a56-91cc-3361f59b57cf-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.546257 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.546975 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49vjn\" (UniqueName: \"kubernetes.io/projected/83b22f4f-f597-4602-a8a2-0ab7b7b580ae-kube-api-access-49vjn\") pod \"etcd-operator-b45778765-7wgpl\" (UID: \"83b22f4f-f597-4602-a8a2-0ab7b7b580ae\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: W1202 13:06:47.559405 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4a1e44b_57f8_49bd_b67f_468593556909.slice/crio-f5bc01a144e58bb5d645406c80c3ecfa135051c97f14bd3ea1853b8533ceaa6a WatchSource:0}: Error finding container f5bc01a144e58bb5d645406c80c3ecfa135051c97f14bd3ea1853b8533ceaa6a: Status 404 returned error can't find the container with id f5bc01a144e58bb5d645406c80c3ecfa135051c97f14bd3ea1853b8533ceaa6a Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.561946 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.562702 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75642\" (UniqueName: \"kubernetes.io/projected/f3877ac3-8496-4b4f-9578-8ee270d55db2-kube-api-access-75642\") pod \"apiserver-7bbb656c7d-x7wfx\" (UID: \"f3877ac3-8496-4b4f-9578-8ee270d55db2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.572597 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.574805 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.574962 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.074938825 +0000 UTC m=+139.031580520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.575259 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.575723 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.075707473 +0000 UTC m=+139.032349178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.596094 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.604281 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q9s7m"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.604563 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.609374 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbxw\" (UniqueName: \"kubernetes.io/projected/24dca536-7872-4a56-91cc-3361f59b57cf-kube-api-access-vxbxw\") pod \"cluster-image-registry-operator-dc59b4c8b-6wjwl\" (UID: \"24dca536-7872-4a56-91cc-3361f59b57cf\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: W1202 13:06:47.613863 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62766e8b_0dd9_42e4_a1f1_4c74ae7f3911.slice/crio-ad2a9c43de628a7e3017e5fa810ae1d4f7692f5cb145aef59d508b39a40ff84a WatchSource:0}: Error finding container ad2a9c43de628a7e3017e5fa810ae1d4f7692f5cb145aef59d508b39a40ff84a: Status 404 returned error can't find the container with id ad2a9c43de628a7e3017e5fa810ae1d4f7692f5cb145aef59d508b39a40ff84a Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.623451 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28fp\" (UniqueName: \"kubernetes.io/projected/6cce0868-2053-4f4c-b9fc-d3e0aaff7de5-kube-api-access-g28fp\") pod \"machine-approver-56656f9798-jjz6n\" (UID: \"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.633153 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7z7tx"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.641035 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6g8l\" (UniqueName: \"kubernetes.io/projected/48f14ad4-c071-4e02-8182-82ffd354c74f-kube-api-access-v6g8l\") pod \"openshift-apiserver-operator-796bbdcf4f-7mdhz\" (UID: \"48f14ad4-c071-4e02-8182-82ffd354c74f\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.660518 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-76gpw"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.665666 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hlts9"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.668361 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghk2k\" (UniqueName: \"kubernetes.io/projected/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-kube-api-access-ghk2k\") pod \"route-controller-manager-6576b87f9c-fqcrt\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.668484 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6fvsl"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.676837 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.677794 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.177773927 +0000 UTC m=+139.134415622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.685821 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.703126 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl82h\" (UniqueName: \"kubernetes.io/projected/9ae02a02-5fff-4fad-9ed7-a18eb58d03b7-kube-api-access-hl82h\") pod \"service-ca-operator-777779d784-knpzb\" (UID: \"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.705678 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.711631 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.751529 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.767116 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6d2x\" (UniqueName: \"kubernetes.io/projected/44387809-9d4a-43c6-b61d-364ab1b6c48f-kube-api-access-x6d2x\") pod \"marketplace-operator-79b997595-ldhmm\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.774124 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7wgpl"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.779756 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.780196 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.280183389 +0000 UTC m=+139.236825084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.784086 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjn4c\" (UniqueName: \"kubernetes.io/projected/18606d46-7784-43be-826b-1b1266255201-kube-api-access-kjn4c\") pod \"multus-admission-controller-857f4d67dd-fwlnc\" (UID: \"18606d46-7784-43be-826b-1b1266255201\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.803392 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79n7q\" (UniqueName: \"kubernetes.io/projected/ea95c612-3884-488c-a153-9c6d32cc54aa-kube-api-access-79n7q\") pod \"machine-config-operator-74547568cd-kkgqq\" (UID: \"ea95c612-3884-488c-a153-9c6d32cc54aa\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.817004 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.827611 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ngk2\" (UniqueName: \"kubernetes.io/projected/2abf7968-024b-4d46-882a-0d8029b116e2-kube-api-access-6ngk2\") pod \"kube-storage-version-migrator-operator-b67b599dd-wx9vk\" (UID: \"2abf7968-024b-4d46-882a-0d8029b116e2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.832332 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.839676 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.841611 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzjv4\" (UniqueName: \"kubernetes.io/projected/233aca62-4621-42ea-b50b-ee87e52af8a9-kube-api-access-qzjv4\") pod \"catalog-operator-68c6474976-ghf6z\" (UID: \"233aca62-4621-42ea-b50b-ee87e52af8a9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.847594 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.848088 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jx7zr"] Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.862330 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2x5f\" (UniqueName: \"kubernetes.io/projected/77e43a6d-f759-4340-a285-46c85b8044cd-kube-api-access-q2x5f\") pod \"machine-config-server-8wm9d\" (UID: \"77e43a6d-f759-4340-a285-46c85b8044cd\") " pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.878320 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.880711 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.880838 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.380798817 +0000 UTC m=+139.337440512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.881034 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.881337 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.38132909 +0000 UTC m=+139.337970775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.883139 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkdb4\" (UniqueName: \"kubernetes.io/projected/6ca9e92f-9c52-4a42-a6b6-e62d2e49e347-kube-api-access-kkdb4\") pod \"packageserver-d55dfcdfc-v7nvs\" (UID: \"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.901431 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62013aee-5ac1-47d6-a5fa-cdadd084ebad-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-qb9tf\" (UID: \"62013aee-5ac1-47d6-a5fa-cdadd084ebad\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.921289 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fg6t\" (UniqueName: \"kubernetes.io/projected/1de2b4c2-30cb-43ea-bf2a-891bea6af79f-kube-api-access-5fg6t\") pod \"csi-hostpathplugin-g49dl\" (UID: \"1de2b4c2-30cb-43ea-bf2a-891bea6af79f\") " pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.923565 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.929707 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" event={"ID":"f4a1e44b-57f8-49bd-b67f-468593556909","Type":"ContainerStarted","Data":"f5bc01a144e58bb5d645406c80c3ecfa135051c97f14bd3ea1853b8533ceaa6a"} Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.930517 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" event={"ID":"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911","Type":"ContainerStarted","Data":"ad2a9c43de628a7e3017e5fa810ae1d4f7692f5cb145aef59d508b39a40ff84a"} Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.931519 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" event={"ID":"b286cd9d-5c21-43f8-a37d-469a65ec0815","Type":"ContainerStarted","Data":"5dae16458b4ba7655e6e805f3976e04fce62c202e029a7cf7e37be568333da7e"} Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.932112 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" event={"ID":"c17fbfab-7855-44b3-a983-22915062793a","Type":"ContainerStarted","Data":"68a71f7b4b74e885dec1e18716d8f7f1fee6c91728ea370b789df719d579579b"} Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.942033 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzqsz\" (UniqueName: \"kubernetes.io/projected/96b0990e-4363-4b28-88fa-dfde258940ac-kube-api-access-rzqsz\") pod \"dns-default-4jfjh\" (UID: \"96b0990e-4363-4b28-88fa-dfde258940ac\") " pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.952514 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.962304 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf4s5\" (UniqueName: \"kubernetes.io/projected/7a128839-ee0f-4a13-b04d-422d4c76f9ee-kube-api-access-bf4s5\") pod \"router-default-5444994796-zdq97\" (UID: \"7a128839-ee0f-4a13-b04d-422d4c76f9ee\") " pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.967113 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.976068 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.981837 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96c99\" (UniqueName: \"kubernetes.io/projected/4d37326e-fdd1-41e0-8401-5f47038356b3-kube-api-access-96c99\") pod \"control-plane-machine-set-operator-78cbb6b69f-bphpt\" (UID: \"4d37326e-fdd1-41e0-8401-5f47038356b3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.982308 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.982449 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.4824184 +0000 UTC m=+139.439060095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.982815 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.983041 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:47 crc kubenswrapper[4725]: E1202 13:06:47.983096 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.483084487 +0000 UTC m=+139.439726182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.984451 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slg75\" (UniqueName: \"kubernetes.io/projected/68f88274-d112-4899-bc9d-e9766a6478d3-kube-api-access-slg75\") pod \"migrator-59844c95c7-hn6hw\" (UID: \"68f88274-d112-4899-bc9d-e9766a6478d3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.988256 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzdzn\" (UniqueName: \"kubernetes.io/projected/1ed82a59-61d2-430c-b819-734c1c2f00b1-kube-api-access-rzdzn\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.988399 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ed82a59-61d2-430c-b819-734c1c2f00b1-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hf9zr\" (UID: \"1ed82a59-61d2-430c-b819-734c1c2f00b1\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.988711 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2btc\" (UniqueName: \"kubernetes.io/projected/b7cb0652-3548-44ab-8031-eb36ff8c7c6b-kube-api-access-w2btc\") pod \"machine-config-controller-84d6567774-vngft\" (UID: \"b7cb0652-3548-44ab-8031-eb36ff8c7c6b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.990958 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" Dec 02 13:06:47 crc kubenswrapper[4725]: W1202 13:06:47.994144 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87c41c9b_efe7_40ba_867c_1fd805f5ff7e.slice/crio-90f24cf3a9174b5ebfe68ac9d153bf00eb25053aab874890fe7b0f1f89300a27 WatchSource:0}: Error finding container 90f24cf3a9174b5ebfe68ac9d153bf00eb25053aab874890fe7b0f1f89300a27: Status 404 returned error can't find the container with id 90f24cf3a9174b5ebfe68ac9d153bf00eb25053aab874890fe7b0f1f89300a27 Dec 02 13:06:47 crc kubenswrapper[4725]: I1202 13:06:47.997146 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.003129 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxtxl\" (UniqueName: \"kubernetes.io/projected/66462d0f-425e-470b-a4e2-94682d534fe9-kube-api-access-hxtxl\") pod \"collect-profiles-29411340-rnxpx\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.014749 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.019113 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.030359 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx286\" (UniqueName: \"kubernetes.io/projected/6db27d13-bcf4-4596-96ba-5f140c9fb38b-kube-api-access-wx286\") pod \"olm-operator-6b444d44fb-7kfq2\" (UID: \"6db27d13-bcf4-4596-96ba-5f140c9fb38b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.045510 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.051442 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.056099 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/04180a2f-4341-498d-9bd0-485967ae26cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hj8b8\" (UID: \"04180a2f-4341-498d-9bd0-485967ae26cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.063699 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dscpw\" (UniqueName: \"kubernetes.io/projected/4bb9a525-a393-4552-83f9-22ceca686349-kube-api-access-dscpw\") pod \"service-ca-9c57cc56f-llqpv\" (UID: \"4bb9a525-a393-4552-83f9-22ceca686349\") " pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.067363 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8wm9d" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.082337 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk4gk\" (UniqueName: \"kubernetes.io/projected/bd10ea21-eff4-4fe0-9474-02d16b2d135d-kube-api-access-kk4gk\") pod \"ingress-canary-2kwpr\" (UID: \"bd10ea21-eff4-4fe0-9474-02d16b2d135d\") " pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.083932 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.084033 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.084434 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.584417332 +0000 UTC m=+139.541059027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.104526 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc2dl\" (UniqueName: \"kubernetes.io/projected/7900355d-cbde-4a76-91ef-9339769983f9-kube-api-access-pc2dl\") pod \"package-server-manager-789f6589d5-pg7x8\" (UID: \"7900355d-cbde-4a76-91ef-9339769983f9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.186060 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.186426 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.686411224 +0000 UTC m=+139.643052919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.218558 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.233824 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.238001 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.258551 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.261206 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.286983 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.287152 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.787126015 +0000 UTC m=+139.743767710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.287283 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.287688 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.787672319 +0000 UTC m=+139.744314074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.288997 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq"] Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.327913 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.337170 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.359896 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z"] Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.360833 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2kwpr" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.381344 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs"] Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.390986 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.391032 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.891013823 +0000 UTC m=+139.847655518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.391384 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.391851 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.891834653 +0000 UTC m=+139.848476348 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.467313 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz"] Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.499582 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.499925 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:48.999909922 +0000 UTC m=+139.956551617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.529243 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.600669 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.600949 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.10093797 +0000 UTC m=+140.057579665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: W1202 13:06:48.682669 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f14ad4_c071_4e02_8182_82ffd354c74f.slice/crio-184a7e321ebb7d650654eff41e37622e77e49328042dd7c2520593bac0208263 WatchSource:0}: Error finding container 184a7e321ebb7d650654eff41e37622e77e49328042dd7c2520593bac0208263: Status 404 returned error can't find the container with id 184a7e321ebb7d650654eff41e37622e77e49328042dd7c2520593bac0208263 Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.701599 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.702407 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.202380259 +0000 UTC m=+140.159021954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: W1202 13:06:48.716401 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod233aca62_4621_42ea_b50b_ee87e52af8a9.slice/crio-f494cbe9c694a51e2122bc81cd5911c6adb43d2b01413d65dee68937ec6ca01c WatchSource:0}: Error finding container f494cbe9c694a51e2122bc81cd5911c6adb43d2b01413d65dee68937ec6ca01c: Status 404 returned error can't find the container with id f494cbe9c694a51e2122bc81cd5911c6adb43d2b01413d65dee68937ec6ca01c Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.803728 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.804039 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.304027182 +0000 UTC m=+140.260668877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.905132 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:48 crc kubenswrapper[4725]: E1202 13:06:48.905427 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.405412959 +0000 UTC m=+140.362054654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.953259 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" event={"ID":"233aca62-4621-42ea-b50b-ee87e52af8a9","Type":"ContainerStarted","Data":"f494cbe9c694a51e2122bc81cd5911c6adb43d2b01413d65dee68937ec6ca01c"} Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.961421 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-knpzb"] Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.961520 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" event={"ID":"e2b66987-71e9-472b-9d08-5e8d89dfe129","Type":"ContainerStarted","Data":"849f7c90a9e82bf57d900459bc8a9e064fbd8369ba53f7d442cc9a617280060b"} Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.961554 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8wm9d" event={"ID":"77e43a6d-f759-4340-a285-46c85b8044cd","Type":"ContainerStarted","Data":"94e19e4bfaa01daafcc308d63f86e60099535708ce325294f05ea5ef0114c211"} Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.969115 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" event={"ID":"ea95c612-3884-488c-a153-9c6d32cc54aa","Type":"ContainerStarted","Data":"afcf145c920562fcdf9c47fda8d2072c968f7fac67aca6a74eb59ce73622efde"} Dec 02 13:06:48 crc kubenswrapper[4725]: I1202 13:06:48.971916 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" event={"ID":"a49031e7-e7da-47c1-b069-a5f129e2794f","Type":"ContainerStarted","Data":"7d6fd84a2a5ea6a21a18688602eb52f181538ffa65ebf56392d64a09dd41247a"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.001477 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" event={"ID":"f4a1e44b-57f8-49bd-b67f-468593556909","Type":"ContainerStarted","Data":"0c41b40102cda4ea944da2238f3c3bf80b9caf56c7f61a5a3db8fd6855df992b"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.008175 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.008831 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.508798935 +0000 UTC m=+140.465440640 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.012152 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" event={"ID":"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936","Type":"ContainerStarted","Data":"bcf982d8faa6da3b23e5cd6402cfa7a7ef6d0e806a83b9808175af5b31695ce0"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.012187 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" event={"ID":"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936","Type":"ContainerStarted","Data":"faf065f27789d9a99279cd8c1d9974b31ce9304e9a1a2338e7996a78d39dee78"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.019109 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" event={"ID":"bb2e34c8-7e96-44c9-a073-87101442f86e","Type":"ContainerStarted","Data":"3b98dea0d95ee5747cfde0f2c47c2c85c656d717d025d5e787ee013c1d5d6d72"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.036062 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" event={"ID":"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5","Type":"ContainerStarted","Data":"e8557cbf511e8e5cee2c19a4f35ffce6a545ec4cfb446298e684e9b28dea59a9"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.058956 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" event={"ID":"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911","Type":"ContainerStarted","Data":"e9291ff65ad0c3223ee71dddffee9aa2e7373a035281fc3a68d9aacd8642b1f8"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.059376 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.083590 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" event={"ID":"48f14ad4-c071-4e02-8182-82ffd354c74f","Type":"ContainerStarted","Data":"184a7e321ebb7d650654eff41e37622e77e49328042dd7c2520593bac0208263"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.105063 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" event={"ID":"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6","Type":"ContainerStarted","Data":"0896e41ee57150b8b1cebecc6816fb976a5edaafac7651826a4d714e4326109a"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.110013 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.111163 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.611144605 +0000 UTC m=+140.567786300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.123371 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.179724 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" event={"ID":"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347","Type":"ContainerStarted","Data":"d4e09f342e62cef342005aed768960d8baba4a7da4cad917683effb3642fb43c"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.227011 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.228429 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.728416988 +0000 UTC m=+140.685058683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.242921 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" event={"ID":"87c41c9b-efe7-40ba-867c-1fd805f5ff7e","Type":"ContainerStarted","Data":"90f24cf3a9174b5ebfe68ac9d153bf00eb25053aab874890fe7b0f1f89300a27"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.331669 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.331992 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.831964417 +0000 UTC m=+140.788606112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422519 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422593 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422617 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" event={"ID":"b286cd9d-5c21-43f8-a37d-469a65ec0815","Type":"ContainerStarted","Data":"9a2386fa1581fe5c09b8eed732717f37bb688fc4f694437ee8d11aa43d105963"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422638 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zdq97" event={"ID":"7a128839-ee0f-4a13-b04d-422d4c76f9ee","Type":"ContainerStarted","Data":"b926bf756a40d4528b49aa95a9719975803ad22704764431bda5203765f4c12b"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422652 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7z7tx" event={"ID":"7971baa9-a25e-47e2-ac21-8c89db535872","Type":"ContainerStarted","Data":"3bab371b695af668c432a4add5d8efe243d633031b9deacdd74b81ed99e0352e"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" event={"ID":"c17fbfab-7855-44b3-a983-22915062793a","Type":"ContainerStarted","Data":"d45f72a729b7caaad6938767c89c0eb72a89df9bd1f1f9e0d462aa2a17e37a19"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.422679 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" event={"ID":"83b22f4f-f597-4602-a8a2-0ab7b7b580ae","Type":"ContainerStarted","Data":"fc8c72d7d9c2d822aa33eba4e3298e94573943c9becc0fe42bd34f3c9f5bfc99"} Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.433504 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.435622 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:49.935607469 +0000 UTC m=+140.892249164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.534969 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.536077 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.036057564 +0000 UTC m=+140.992699259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.638526 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.638902 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.138887866 +0000 UTC m=+141.095529561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.657525 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" podStartSLOduration=121.657505907 podStartE2EDuration="2m1.657505907s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:49.655917429 +0000 UTC m=+140.612559124" watchObservedRunningTime="2025-12-02 13:06:49.657505907 +0000 UTC m=+140.614147602" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.657909 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7z7tx" podStartSLOduration=121.657903587 podStartE2EDuration="2m1.657903587s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:49.634847838 +0000 UTC m=+140.591489533" watchObservedRunningTime="2025-12-02 13:06:49.657903587 +0000 UTC m=+140.614545282" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.740220 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.740667 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.240642542 +0000 UTC m=+141.197284247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.741001 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.741375 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.241361569 +0000 UTC m=+141.198003264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.763684 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pb5sr" podStartSLOduration=121.76366544 podStartE2EDuration="2m1.76366544s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:49.695317993 +0000 UTC m=+140.651959688" watchObservedRunningTime="2025-12-02 13:06:49.76366544 +0000 UTC m=+140.720307135" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.819664 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-hgcrs" podStartSLOduration=121.819645287 podStartE2EDuration="2m1.819645287s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:49.815923627 +0000 UTC m=+140.772565332" watchObservedRunningTime="2025-12-02 13:06:49.819645287 +0000 UTC m=+140.776286982" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.840971 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" podStartSLOduration=121.840952753 podStartE2EDuration="2m1.840952753s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:49.840028051 +0000 UTC m=+140.796669756" watchObservedRunningTime="2025-12-02 13:06:49.840952753 +0000 UTC m=+140.797594448" Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.841606 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.843742 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.34371541 +0000 UTC m=+141.300357105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:49 crc kubenswrapper[4725]: I1202 13:06:49.967179 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:49 crc kubenswrapper[4725]: E1202 13:06:49.967554 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.467532901 +0000 UTC m=+141.424174596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.356570 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.356679 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.856656232 +0000 UTC m=+141.813297927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.356984 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.357318 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.857305758 +0000 UTC m=+141.813947453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.365355 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-h8fhj" event={"ID":"c17fbfab-7855-44b3-a983-22915062793a","Type":"ContainerStarted","Data":"618f2b4eac7e277ae140bd244e0fdb8ecbea27f30cb33075450a9aa4afdb630e"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.368883 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" event={"ID":"e2b66987-71e9-472b-9d08-5e8d89dfe129","Type":"ContainerStarted","Data":"ec50b525b49371f63bf9ac5c9dbac6c9a5be5e6d3b4074284ab8ab434e3ed530"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.370522 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zdq97" event={"ID":"7a128839-ee0f-4a13-b04d-422d4c76f9ee","Type":"ContainerStarted","Data":"8c4b91887c3fd74dc27c061ee5c2967a01dd20e5912dcc195259e058af1264ad"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.378517 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" event={"ID":"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6","Type":"ContainerStarted","Data":"67f4cff7e438a5285df7eef3f7fe0ba6597347bc774ed1574e28593357926d13"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.379616 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" event={"ID":"a49031e7-e7da-47c1-b069-a5f129e2794f","Type":"ContainerStarted","Data":"59aa628894b9ad189fac27d47cb2ccf2ba65b48c053b9bf67eba0df9c05c176d"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.383109 4725 generic.go:334] "Generic (PLEG): container finished" podID="87c41c9b-efe7-40ba-867c-1fd805f5ff7e" containerID="df541dfee8fe2b23ff33c5186c192539cd432e50fd934a555a3c0d1ded48d614" exitCode=0 Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.383174 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" event={"ID":"87c41c9b-efe7-40ba-867c-1fd805f5ff7e","Type":"ContainerDied","Data":"df541dfee8fe2b23ff33c5186c192539cd432e50fd934a555a3c0d1ded48d614"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.385091 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" event={"ID":"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7","Type":"ContainerStarted","Data":"6d5ed65b5416f5b6c41aa5ade14d6a15c331e54cb1016c8998d1d6dc122b0b35"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.393419 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7z7tx" event={"ID":"7971baa9-a25e-47e2-ac21-8c89db535872","Type":"ContainerStarted","Data":"0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.406119 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f4r68" podStartSLOduration=122.40609763 podStartE2EDuration="2m2.40609763s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:50.403926007 +0000 UTC m=+141.360567712" watchObservedRunningTime="2025-12-02 13:06:50.40609763 +0000 UTC m=+141.362739315" Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.410326 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" event={"ID":"ea95c612-3884-488c-a153-9c6d32cc54aa","Type":"ContainerStarted","Data":"d0b6056a41ef8b6b1c656ea0fe2376ac7757f1d2b9c24d32bb2b22b40de0f8cb"} Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.432596 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-hlts9" podStartSLOduration=122.432577612 podStartE2EDuration="2m2.432577612s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:50.432347216 +0000 UTC m=+141.388988931" watchObservedRunningTime="2025-12-02 13:06:50.432577612 +0000 UTC m=+141.389219307" Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.456655 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zdq97" podStartSLOduration=122.45663600500001 podStartE2EDuration="2m2.456636005s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:50.453666663 +0000 UTC m=+141.410308378" watchObservedRunningTime="2025-12-02 13:06:50.456636005 +0000 UTC m=+141.413277700" Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.457904 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.457996 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.957980948 +0000 UTC m=+141.914622643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.458560 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.459753 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:50.95974117 +0000 UTC m=+141.916382865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.561608 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.562809 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.062793828 +0000 UTC m=+142.019435523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.662671 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.662999 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.162987566 +0000 UTC m=+142.119629261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.765187 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.766505 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.266489805 +0000 UTC m=+142.223131500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.869652 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.870032 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.370016114 +0000 UTC m=+142.326657809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.909250 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wst96"] Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.948561 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx"] Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.966559 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx"] Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.970897 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:50 crc kubenswrapper[4725]: E1202 13:06:50.971519 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.471497953 +0000 UTC m=+142.428139658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:50 crc kubenswrapper[4725]: I1202 13:06:50.976520 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:50.996804 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt"] Dec 02 13:06:51 crc kubenswrapper[4725]: W1202 13:06:51.037594 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d37326e_fdd1_41e0_8401_5f47038356b3.slice/crio-e11fe6f7078a04e333ee7ff574eb0659cdd4fc3a75cf266e08a06d6d3f635e5c WatchSource:0}: Error finding container e11fe6f7078a04e333ee7ff574eb0659cdd4fc3a75cf266e08a06d6d3f635e5c: Status 404 returned error can't find the container with id e11fe6f7078a04e333ee7ff574eb0659cdd4fc3a75cf266e08a06d6d3f635e5c Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.047554 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.054340 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:51 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:51 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:51 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.054389 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.073377 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.073737 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.573724781 +0000 UTC m=+142.530366476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.115924 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gp8pw"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.124415 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fwlnc"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.177060 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.177495 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.677453065 +0000 UTC m=+142.634094760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.200786 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk"] Dec 02 13:06:51 crc kubenswrapper[4725]: W1202 13:06:51.210085 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18606d46_7784_43be_826b_1b1266255201.slice/crio-71d61b6f541b8c4b6b3e3d31e7a62a0e2ab6e5ac3d37b7b4e6e614376e659c62 WatchSource:0}: Error finding container 71d61b6f541b8c4b6b3e3d31e7a62a0e2ab6e5ac3d37b7b4e6e614376e659c62: Status 404 returned error can't find the container with id 71d61b6f541b8c4b6b3e3d31e7a62a0e2ab6e5ac3d37b7b4e6e614376e659c62 Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.278325 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.278852 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.778839822 +0000 UTC m=+142.735481517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.382994 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.383324 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.883307744 +0000 UTC m=+142.839949439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.417500 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-llqpv"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.480235 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2kwpr"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.486928 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.487273 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:51.987260743 +0000 UTC m=+142.943902438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.488289 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.542932 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" event={"ID":"ea95c612-3884-488c-a153-9c6d32cc54aa","Type":"ContainerStarted","Data":"d1aed135b9e797fa8e0febc0e379ff231e080f7f7c224a583f88d6d0ab36dd36"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.545559 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4jfjh"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.546645 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldhmm"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.580444 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vngft"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.594012 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.594802 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8"] Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.595109 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.095078347 +0000 UTC m=+143.051720092 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.604587 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.630841 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" event={"ID":"233aca62-4621-42ea-b50b-ee87e52af8a9","Type":"ContainerStarted","Data":"5241e6d7712751d880b9d82d6da630de126a18a7817a8f067071f1702a7ac273"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.642314 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-g49dl"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.642373 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.642395 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.654960 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.672753 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" event={"ID":"4d37326e-fdd1-41e0-8401-5f47038356b3","Type":"ContainerStarted","Data":"e11fe6f7078a04e333ee7ff574eb0659cdd4fc3a75cf266e08a06d6d3f635e5c"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.675710 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.675810 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kkgqq" podStartSLOduration=123.675794293 podStartE2EDuration="2m3.675794293s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:51.59480982 +0000 UTC m=+142.551451515" watchObservedRunningTime="2025-12-02 13:06:51.675794293 +0000 UTC m=+142.632435988" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.698209 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.699875 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.199861356 +0000 UTC m=+143.156503051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.704535 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.704883 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr"] Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.726680 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" event={"ID":"f3877ac3-8496-4b4f-9578-8ee270d55db2","Type":"ContainerStarted","Data":"6db22a0c9481d1fc191c71898e097e19d369b4c3306c4db1350a03b9db59a2d6"} Dec 02 13:06:51 crc kubenswrapper[4725]: W1202 13:06:51.746593 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7900355d_cbde_4a76_91ef_9339769983f9.slice/crio-c2a58a134e390238c7e7bb18b1cc3f2393f56c883799545d934774ef8b0aa548 WatchSource:0}: Error finding container c2a58a134e390238c7e7bb18b1cc3f2393f56c883799545d934774ef8b0aa548: Status 404 returned error can't find the container with id c2a58a134e390238c7e7bb18b1cc3f2393f56c883799545d934774ef8b0aa548 Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.775907 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ghf6z" podStartSLOduration=123.775889119 podStartE2EDuration="2m3.775889119s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:51.724040032 +0000 UTC m=+142.680681747" watchObservedRunningTime="2025-12-02 13:06:51.775889119 +0000 UTC m=+142.732530814" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.801915 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.802846 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.302825892 +0000 UTC m=+143.259467597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: W1202 13:06:51.817857 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62013aee_5ac1_47d6_a5fa_cdadd084ebad.slice/crio-e0962c5afffe4ec6f006b4e5e18b228fe074773373a1f473361bcf23a59e5f10 WatchSource:0}: Error finding container e0962c5afffe4ec6f006b4e5e18b228fe074773373a1f473361bcf23a59e5f10: Status 404 returned error can't find the container with id e0962c5afffe4ec6f006b4e5e18b228fe074773373a1f473361bcf23a59e5f10 Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.822169 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" event={"ID":"9f8a2bd4-3ba5-49ea-8c4d-c3aedeb69936","Type":"ContainerStarted","Data":"73fe911eb7f5a16ed0492661f2dbef2ef9c0e89f6d0236f7b1b3c791d5bec59b"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.852406 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" event={"ID":"48f14ad4-c071-4e02-8182-82ffd354c74f","Type":"ContainerStarted","Data":"601807d11775774920803c95a35435566ab5d4f4aa246217c17fd0e42d4f8a8b"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.898851 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mdhz" podStartSLOduration=123.898837529 podStartE2EDuration="2m3.898837529s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:51.897968207 +0000 UTC m=+142.854609892" watchObservedRunningTime="2025-12-02 13:06:51.898837529 +0000 UTC m=+142.855479224" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.900039 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c89qx" podStartSLOduration=123.900030817 podStartE2EDuration="2m3.900030817s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:51.877822339 +0000 UTC m=+142.834464044" watchObservedRunningTime="2025-12-02 13:06:51.900030817 +0000 UTC m=+142.856672512" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.900384 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" event={"ID":"e4bd2696-bbc8-4e29-b6d5-cec4ab717cf6","Type":"ContainerStarted","Data":"34f2b4bb98d388b2f752ad5b88e4d62b50dbd5e58157ed2b3e1d29b876bade18"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.905884 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:51 crc kubenswrapper[4725]: E1202 13:06:51.906886 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.406874803 +0000 UTC m=+143.363516498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.914567 4725 generic.go:334] "Generic (PLEG): container finished" podID="bb2e34c8-7e96-44c9-a073-87101442f86e" containerID="d8302ae6ecd16e9b4a2bf73458c4be55ec3484237070e4754f3f6f4e471a14cd" exitCode=0 Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.914644 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" event={"ID":"bb2e34c8-7e96-44c9-a073-87101442f86e","Type":"ContainerDied","Data":"d8302ae6ecd16e9b4a2bf73458c4be55ec3484237070e4754f3f6f4e471a14cd"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.930629 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" event={"ID":"9ae02a02-5fff-4fad-9ed7-a18eb58d03b7","Type":"ContainerStarted","Data":"6a5e39f87bffcfa70f44ecf1c90dd83f6617eff7db3e6972bda04fe028919c3b"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.956119 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" event={"ID":"2abf7968-024b-4d46-882a-0d8029b116e2","Type":"ContainerStarted","Data":"858374690ae123b29a485cd91e9e3c9b4dbca93f3aca042922e020c1c8683859"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.964735 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" event={"ID":"66462d0f-425e-470b-a4e2-94682d534fe9","Type":"ContainerStarted","Data":"3a0e3a76fff0d396e51bf96e1542597b76841b39cadaefda8908c7044594e915"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.964789 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" event={"ID":"66462d0f-425e-470b-a4e2-94682d534fe9","Type":"ContainerStarted","Data":"534d3005718794516154251dc7af2d3162fd58c1f9c1ce9efba52faef1e12367"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.972611 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" event={"ID":"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5","Type":"ContainerStarted","Data":"75fc0372696ecc7c5a5477cfe133e271615e98e4dc6d6ef7adf6343e8902f974"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.980194 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" event={"ID":"87c41c9b-efe7-40ba-867c-1fd805f5ff7e","Type":"ContainerStarted","Data":"e681a121bc5866007f33824b05a376e26d3c41ba662b54a4a79ed7594868f407"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.980299 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.991904 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jx7zr" podStartSLOduration=123.991881674 podStartE2EDuration="2m3.991881674s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:51.948928873 +0000 UTC m=+142.905570568" watchObservedRunningTime="2025-12-02 13:06:51.991881674 +0000 UTC m=+142.948523369" Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.998102 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" event={"ID":"2034b4c1-5d0f-42d4-951d-afd68a259ee1","Type":"ContainerStarted","Data":"a4d2f7e538f296ab79e03846aa897c83a86dbf9b3afc25691eff6aaf2fdbdf10"} Dec 02 13:06:51 crc kubenswrapper[4725]: I1202 13:06:51.998524 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.008223 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.031559 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.531535015 +0000 UTC m=+143.488176710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.035524 4725 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wst96 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.26:6443/healthz\": dial tcp 10.217.0.26:6443: connect: connection refused" start-of-body= Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.035584 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" podUID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.26:6443/healthz\": dial tcp 10.217.0.26:6443: connect: connection refused" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.036122 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gp8pw" event={"ID":"a81f9c3a-432a-4454-95fa-aea3a8cc3344","Type":"ContainerStarted","Data":"ee8a0855da6b19bc030d87e2e84f3bbfe0942765ef5d998f2e3f63018a15149c"} Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.036915 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.047496 4725 patch_prober.go:28] interesting pod/downloads-7954f5f757-gp8pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.047565 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gp8pw" podUID="a81f9c3a-432a-4454-95fa-aea3a8cc3344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.056069 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:52 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:52 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:52 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.056149 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.063231 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-knpzb" podStartSLOduration=124.063208452 podStartE2EDuration="2m4.063208452s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.033299007 +0000 UTC m=+142.989940702" watchObservedRunningTime="2025-12-02 13:06:52.063208452 +0000 UTC m=+143.019850147" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.068415 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" event={"ID":"18606d46-7784-43be-826b-1b1266255201","Type":"ContainerStarted","Data":"71d61b6f541b8c4b6b3e3d31e7a62a0e2ab6e5ac3d37b7b4e6e614376e659c62"} Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.093409 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" podStartSLOduration=124.093381324 podStartE2EDuration="2m4.093381324s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.062541936 +0000 UTC m=+143.019183631" watchObservedRunningTime="2025-12-02 13:06:52.093381324 +0000 UTC m=+143.050023019" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.095092 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" podStartSLOduration=124.095078005 podStartE2EDuration="2m4.095078005s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.093922897 +0000 UTC m=+143.050564592" watchObservedRunningTime="2025-12-02 13:06:52.095078005 +0000 UTC m=+143.051719700" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.110095 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.110400 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.610385336 +0000 UTC m=+143.567027031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.113019 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" event={"ID":"83b22f4f-f597-4602-a8a2-0ab7b7b580ae","Type":"ContainerStarted","Data":"1ce0a0c8f5c1f103b51773319904491c93d0d87cd50ce7ee5c4fd72223831b6e"} Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.132359 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8wm9d" event={"ID":"77e43a6d-f759-4340-a285-46c85b8044cd","Type":"ContainerStarted","Data":"fc0014df1417f88e2e09d6a7e386a07b8f3b7d4280fd4bf1c6af4ca13049e88d"} Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.152683 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" podStartSLOduration=124.152666279 podStartE2EDuration="2m4.152666279s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.148765465 +0000 UTC m=+143.105407180" watchObservedRunningTime="2025-12-02 13:06:52.152666279 +0000 UTC m=+143.109307974" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.153775 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" podStartSLOduration=124.153769716 podStartE2EDuration="2m4.153769716s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.117026066 +0000 UTC m=+143.073667761" watchObservedRunningTime="2025-12-02 13:06:52.153769716 +0000 UTC m=+143.110411401" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.189629 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" podStartSLOduration=124.189613045 podStartE2EDuration="2m4.189613045s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.186895639 +0000 UTC m=+143.143537344" watchObservedRunningTime="2025-12-02 13:06:52.189613045 +0000 UTC m=+143.146254740" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.195675 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" event={"ID":"6ca9e92f-9c52-4a42-a6b6-e62d2e49e347","Type":"ContainerStarted","Data":"da284d80f28a88eca75bdf53f5c4f01bbc55e86ed6a5ee22fef8d7453d42cff9"} Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.196148 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.213165 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7wgpl" podStartSLOduration=124.213149515 podStartE2EDuration="2m4.213149515s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.211932266 +0000 UTC m=+143.168573951" watchObservedRunningTime="2025-12-02 13:06:52.213149515 +0000 UTC m=+143.169791210" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.214989 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.714974119 +0000 UTC m=+143.671615814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.215627 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.215920 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.218249 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.718234118 +0000 UTC m=+143.674875803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.227017 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" event={"ID":"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f","Type":"ContainerStarted","Data":"1b7ee3727b445416530c7dd0e6df25c401310e00b705e0708e2742e260ee0ff8"} Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.227057 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.293116 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-gp8pw" podStartSLOduration=124.293098193 podStartE2EDuration="2m4.293098193s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.26779258 +0000 UTC m=+143.224434275" watchObservedRunningTime="2025-12-02 13:06:52.293098193 +0000 UTC m=+143.249739888" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.316491 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" podStartSLOduration=124.316470219 podStartE2EDuration="2m4.316470219s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.315565637 +0000 UTC m=+143.272207332" watchObservedRunningTime="2025-12-02 13:06:52.316470219 +0000 UTC m=+143.273111914" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.318356 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8wm9d" podStartSLOduration=7.318348445 podStartE2EDuration="7.318348445s" podCreationTimestamp="2025-12-02 13:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.293725118 +0000 UTC m=+143.250366813" watchObservedRunningTime="2025-12-02 13:06:52.318348445 +0000 UTC m=+143.274990140" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.335077 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.336767 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.836750941 +0000 UTC m=+143.793392636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.339657 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.345642 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.845623396 +0000 UTC m=+143.802265161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.363947 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v7nvs" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.372406 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" podStartSLOduration=124.372390745 podStartE2EDuration="2m4.372390745s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:52.370094679 +0000 UTC m=+143.326736374" watchObservedRunningTime="2025-12-02 13:06:52.372390745 +0000 UTC m=+143.329032440" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.448946 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.450241 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:52.950220911 +0000 UTC m=+143.906862606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.507282 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.552089 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.552438 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.052426788 +0000 UTC m=+144.009068483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.653388 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.654058 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.154037671 +0000 UTC m=+144.110679376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.755038 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.755354 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.255342426 +0000 UTC m=+144.211984121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.856801 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.857029 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.35699292 +0000 UTC m=+144.313634615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.857271 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.857613 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.357606805 +0000 UTC m=+144.314248500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.951035 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w4rh7"] Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.952502 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.954618 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.958330 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:52 crc kubenswrapper[4725]: E1202 13:06:52.958767 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.458749016 +0000 UTC m=+144.415390711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:52 crc kubenswrapper[4725]: I1202 13:06:52.970970 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w4rh7"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.056635 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:53 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:53 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:53 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.056718 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.060444 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.060504 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-utilities\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.060525 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-catalog-content\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.060605 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl5lr\" (UniqueName: \"kubernetes.io/projected/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-kube-api-access-fl5lr\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.060879 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.560868221 +0000 UTC m=+144.517509916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.137627 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ntm7v"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.138488 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.154514 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.163131 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.163502 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-utilities\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.163540 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-catalog-content\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.163650 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl5lr\" (UniqueName: \"kubernetes.io/projected/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-kube-api-access-fl5lr\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.164241 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.664215486 +0000 UTC m=+144.620857191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.164775 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-utilities\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.164825 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-catalog-content\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.200565 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntm7v"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.214365 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl5lr\" (UniqueName: \"kubernetes.io/projected/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-kube-api-access-fl5lr\") pod \"certified-operators-w4rh7\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.264274 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" event={"ID":"6db27d13-bcf4-4596-96ba-5f140c9fb38b","Type":"ContainerStarted","Data":"68e21f444c375c65f3220c660f8b21d077a1d5f580746ade07adab7c76c53bff"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.264592 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" event={"ID":"6db27d13-bcf4-4596-96ba-5f140c9fb38b","Type":"ContainerStarted","Data":"bedcd1a1bbaba8d62c517a22c77c1389878ec19518f708f5b7123b0267a5500e"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.267050 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.268617 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.268657 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmvm6\" (UniqueName: \"kubernetes.io/projected/362e7a06-0d49-46c4-b7e9-1bfc52125374-kube-api-access-zmvm6\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.268699 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-utilities\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.268757 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-catalog-content\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.269127 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.769115558 +0000 UTC m=+144.725757253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.288506 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gp8pw" event={"ID":"a81f9c3a-432a-4454-95fa-aea3a8cc3344","Type":"ContainerStarted","Data":"0065db2adbd24a1d03ee17ce51794259994c7b3f464d9408f7a345b7a108b698"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.289370 4725 patch_prober.go:28] interesting pod/downloads-7954f5f757-gp8pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.289407 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gp8pw" podUID="a81f9c3a-432a-4454-95fa-aea3a8cc3344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.302558 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.303332 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7kfq2" podStartSLOduration=125.303323557 podStartE2EDuration="2m5.303323557s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.303296516 +0000 UTC m=+144.259938211" watchObservedRunningTime="2025-12-02 13:06:53.303323557 +0000 UTC m=+144.259965252" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.304620 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" event={"ID":"4d37326e-fdd1-41e0-8401-5f47038356b3","Type":"ContainerStarted","Data":"a19ba040be92362945732066e694ae0cba6d698cab8d56ba9c602fc38f60a87d"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.325897 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4jfjh" event={"ID":"96b0990e-4363-4b28-88fa-dfde258940ac","Type":"ContainerStarted","Data":"9f5b92b075a52ae84c9fefa1d9e1032a0c5850df1a0991acc596dba1f9bfec2b"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.325939 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4jfjh" event={"ID":"96b0990e-4363-4b28-88fa-dfde258940ac","Type":"ContainerStarted","Data":"76e70ce81fe1804f7cab97a9689889777fe49d2e5f908511cc64663bb7523fcf"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.341824 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-wx9vk" event={"ID":"2abf7968-024b-4d46-882a-0d8029b116e2","Type":"ContainerStarted","Data":"51c8e0a6826328e1a738e6ddc78f5e06a7ce3079b48ccc643d732cc3e5ccfb30"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.369597 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.369765 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-utilities\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.369871 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-catalog-content\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.369944 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmvm6\" (UniqueName: \"kubernetes.io/projected/362e7a06-0d49-46c4-b7e9-1bfc52125374-kube-api-access-zmvm6\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.370545 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.870529596 +0000 UTC m=+144.827171291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.370898 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-utilities\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.372345 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-catalog-content\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.372972 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9lz9d"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.374316 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.376586 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.384536 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" event={"ID":"1ed82a59-61d2-430c-b819-734c1c2f00b1","Type":"ContainerStarted","Data":"c59be62f97f166c2dc26dad6ca585f7d151242a7f5388813953146f91670cb59"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.384573 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" event={"ID":"1ed82a59-61d2-430c-b819-734c1c2f00b1","Type":"ContainerStarted","Data":"509839a74a656995eb72159b8f2fae5bae12b7d314bdaaf9342a93253c691ae4"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.384588 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" event={"ID":"1ed82a59-61d2-430c-b819-734c1c2f00b1","Type":"ContainerStarted","Data":"6e26fee6aca4bc70d16bde9670d4fecd7f42c13eb0f58de466ff131c296e4e81"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.419582 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lz9d"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.421238 4725 generic.go:334] "Generic (PLEG): container finished" podID="f3877ac3-8496-4b4f-9578-8ee270d55db2" containerID="cfde77bcb38dc35b9832ebb3d5b3f61472238257c12f0d81a30af1cb82c5027e" exitCode=0 Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.421289 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" event={"ID":"f3877ac3-8496-4b4f-9578-8ee270d55db2","Type":"ContainerDied","Data":"cfde77bcb38dc35b9832ebb3d5b3f61472238257c12f0d81a30af1cb82c5027e"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.423864 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmvm6\" (UniqueName: \"kubernetes.io/projected/362e7a06-0d49-46c4-b7e9-1bfc52125374-kube-api-access-zmvm6\") pod \"community-operators-ntm7v\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.423928 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" event={"ID":"b7cb0652-3548-44ab-8031-eb36ff8c7c6b","Type":"ContainerStarted","Data":"de71ad3eecc88594f629ed4f7948e368cb9801d868ff3581e7003be22622e74e"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.423952 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" event={"ID":"b7cb0652-3548-44ab-8031-eb36ff8c7c6b","Type":"ContainerStarted","Data":"8ad58a7d7aff4c9994bdf52bb5b3789c742ef1eb45e13a5877e67796fedc95dc"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.423960 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" event={"ID":"b7cb0652-3548-44ab-8031-eb36ff8c7c6b","Type":"ContainerStarted","Data":"f0120c447ec01e6c64bfcd7182283a0d50460c74f255febf84b12d1f6be40987"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.427180 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bphpt" podStartSLOduration=125.427155808 podStartE2EDuration="2m5.427155808s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.418204161 +0000 UTC m=+144.374845856" watchObservedRunningTime="2025-12-02 13:06:53.427155808 +0000 UTC m=+144.383797503" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.447716 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" event={"ID":"68f88274-d112-4899-bc9d-e9766a6478d3","Type":"ContainerStarted","Data":"e609e0d5553a2be6b07c8d2e248539f33b56842400463aab59003b3f235f093c"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.447762 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" event={"ID":"68f88274-d112-4899-bc9d-e9766a6478d3","Type":"ContainerStarted","Data":"abfcb695ccdd63e80bdb2202fe5481a0819883395f41b21abba3ef83d8c4648c"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.447772 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" event={"ID":"68f88274-d112-4899-bc9d-e9766a6478d3","Type":"ContainerStarted","Data":"3759ba3d4c698649c27bde978dfcf407445a2c7df386fc57b55d3c0ce0bf0846"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.472197 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qps46\" (UniqueName: \"kubernetes.io/projected/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-kube-api-access-qps46\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.472231 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-utilities\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.472332 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-catalog-content\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.472356 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.475154 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:53.975141661 +0000 UTC m=+144.931783356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.483733 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.492728 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" event={"ID":"24dca536-7872-4a56-91cc-3361f59b57cf","Type":"ContainerStarted","Data":"ccf71494cf2e55876d676639b00f19c753869fdeab8e4fc06d3d9e2cefc88a75"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.492808 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" event={"ID":"24dca536-7872-4a56-91cc-3361f59b57cf","Type":"ContainerStarted","Data":"2d29e49c8981f9af68c8673fa22e633c713538dcb1394ce23a6d5651963d3c5c"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.494638 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hf9zr" podStartSLOduration=125.494624013 podStartE2EDuration="2m5.494624013s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.49408939 +0000 UTC m=+144.450731085" watchObservedRunningTime="2025-12-02 13:06:53.494624013 +0000 UTC m=+144.451265698" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.540884 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" event={"ID":"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f","Type":"ContainerStarted","Data":"12fd158fb839f459bde6993a22ca260a336cbb67a03efc2cd8c905805c7fb705"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.563793 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-brw47"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.564713 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.574580 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.574879 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-catalog-content\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.574980 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-utilities\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.575001 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qps46\" (UniqueName: \"kubernetes.io/projected/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-kube-api-access-qps46\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.575511 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.075495803 +0000 UTC m=+145.032137498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.577696 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-catalog-content\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.578246 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-utilities\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.582851 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" event={"ID":"04180a2f-4341-498d-9bd0-485967ae26cd","Type":"ContainerStarted","Data":"42e43e178dd2b61e6a5a169b966ec0bf74750acff97423116f9091e9b377c557"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.582892 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" event={"ID":"04180a2f-4341-498d-9bd0-485967ae26cd","Type":"ContainerStarted","Data":"3028b5442b45dcbaa3f919cdf07c479d064cac7211431b92f0370abcc4400e37"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.593315 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brw47"] Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.595061 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6wjwl" podStartSLOduration=125.595041877 podStartE2EDuration="2m5.595041877s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.583679062 +0000 UTC m=+144.540320757" watchObservedRunningTime="2025-12-02 13:06:53.595041877 +0000 UTC m=+144.551683572" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.640944 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qps46\" (UniqueName: \"kubernetes.io/projected/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-kube-api-access-qps46\") pod \"certified-operators-9lz9d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.647784 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hn6hw" podStartSLOduration=125.647768425 podStartE2EDuration="2m5.647768425s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.617487281 +0000 UTC m=+144.574128966" watchObservedRunningTime="2025-12-02 13:06:53.647768425 +0000 UTC m=+144.604410120" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.677632 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwss2\" (UniqueName: \"kubernetes.io/projected/7044a434-25d2-4aed-bcd1-17864366d235-kube-api-access-nwss2\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.677685 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-utilities\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.677799 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-catalog-content\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.677835 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.678195 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.178180392 +0000 UTC m=+145.134822087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.682719 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" event={"ID":"4bb9a525-a393-4552-83f9-22ceca686349","Type":"ContainerStarted","Data":"e8c70fb30dd971f99e5d1fc8a3f0d13ffd4e486728160eacd10374b84afd719d"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.682820 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" event={"ID":"4bb9a525-a393-4552-83f9-22ceca686349","Type":"ContainerStarted","Data":"da4712bb6647e1ed76b989be5e5a05f50e2ee42f8fda02996b442ad7b8711c2b"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.688888 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vngft" podStartSLOduration=125.688869861 podStartE2EDuration="2m5.688869861s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.648539703 +0000 UTC m=+144.605181398" watchObservedRunningTime="2025-12-02 13:06:53.688869861 +0000 UTC m=+144.645511556" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.703071 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.708668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" event={"ID":"62013aee-5ac1-47d6-a5fa-cdadd084ebad","Type":"ContainerStarted","Data":"c408c40ae54bb95790a68e28cd86b5dcedbb4391c81351ccbb4ef089b2e694b8"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.708721 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" event={"ID":"62013aee-5ac1-47d6-a5fa-cdadd084ebad","Type":"ContainerStarted","Data":"e0962c5afffe4ec6f006b4e5e18b228fe074773373a1f473361bcf23a59e5f10"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.735839 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" event={"ID":"7900355d-cbde-4a76-91ef-9339769983f9","Type":"ContainerStarted","Data":"8768e3f5f964796639f61eb2c311784befbbe2110d6d32519bfcf910e6a8f5a3"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.735883 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" event={"ID":"7900355d-cbde-4a76-91ef-9339769983f9","Type":"ContainerStarted","Data":"a5e5b87d9fe4292763945d6bf5dc1b30e9a626754415358597b2942d21c79dac"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.735892 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" event={"ID":"7900355d-cbde-4a76-91ef-9339769983f9","Type":"ContainerStarted","Data":"c2a58a134e390238c7e7bb18b1cc3f2393f56c883799545d934774ef8b0aa548"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.736565 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.750055 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jjz6n" event={"ID":"6cce0868-2053-4f4c-b9fc-d3e0aaff7de5","Type":"ContainerStarted","Data":"cd85470ddcc65e2b96c7d8cf0379e111fd672337605bc7476ad0294212c6a721"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.755749 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" event={"ID":"44387809-9d4a-43c6-b61d-364ab1b6c48f","Type":"ContainerStarted","Data":"edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.755786 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" event={"ID":"44387809-9d4a-43c6-b61d-364ab1b6c48f","Type":"ContainerStarted","Data":"157d980855f44a64e8f9d0f0e12bebacef5380aff82d60d466efa349363390e5"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.756621 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.761523 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hj8b8" podStartSLOduration=125.761504751 podStartE2EDuration="2m5.761504751s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.754725717 +0000 UTC m=+144.711367412" watchObservedRunningTime="2025-12-02 13:06:53.761504751 +0000 UTC m=+144.718146446" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.764649 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" event={"ID":"2034b4c1-5d0f-42d4-951d-afd68a259ee1","Type":"ContainerStarted","Data":"78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.780413 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.780662 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-catalog-content\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.780956 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwss2\" (UniqueName: \"kubernetes.io/projected/7044a434-25d2-4aed-bcd1-17864366d235-kube-api-access-nwss2\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.781001 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-utilities\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.782065 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.282049579 +0000 UTC m=+145.238691274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.784157 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-catalog-content\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.799064 4725 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ldhmm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.799122 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.806021 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-utilities\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.819886 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-llqpv" podStartSLOduration=125.819860226 podStartE2EDuration="2m5.819860226s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.806956403 +0000 UTC m=+144.763598088" watchObservedRunningTime="2025-12-02 13:06:53.819860226 +0000 UTC m=+144.776501921" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.864230 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" event={"ID":"1de2b4c2-30cb-43ea-bf2a-891bea6af79f","Type":"ContainerStarted","Data":"6d8ca847b999b3a3a6e71bba897c988be3c0dc824c4923b2dd5057f53ae387c6"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.868782 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" podStartSLOduration=125.868763681 podStartE2EDuration="2m5.868763681s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.863054442 +0000 UTC m=+144.819696137" watchObservedRunningTime="2025-12-02 13:06:53.868763681 +0000 UTC m=+144.825405366" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.869879 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2kwpr" event={"ID":"bd10ea21-eff4-4fe0-9474-02d16b2d135d","Type":"ContainerStarted","Data":"92406e5f8c8fe87e689e50f734226f99eba6141b6003745d01e57f03794e4e17"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.869914 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2kwpr" event={"ID":"bd10ea21-eff4-4fe0-9474-02d16b2d135d","Type":"ContainerStarted","Data":"23444ce41add7bc3a05033a146b031c5c967f3bcbcbc98765b1baa052620b962"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.870782 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwss2\" (UniqueName: \"kubernetes.io/projected/7044a434-25d2-4aed-bcd1-17864366d235-kube-api-access-nwss2\") pod \"community-operators-brw47\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.887868 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.888293 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.388280834 +0000 UTC m=+145.344922529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.921707 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" event={"ID":"18606d46-7784-43be-826b-1b1266255201","Type":"ContainerStarted","Data":"dd3a634fdecca3d0c7e94ff42ead3d78a8e766e4cde6afff4ee4dac156eed05f"} Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.928203 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.929709 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.943924 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" podStartSLOduration=125.943902582 podStartE2EDuration="2m5.943902582s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.905102952 +0000 UTC m=+144.861744647" watchObservedRunningTime="2025-12-02 13:06:53.943902582 +0000 UTC m=+144.900544277" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.944064 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-qb9tf" podStartSLOduration=125.944060366 podStartE2EDuration="2m5.944060366s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:53.939989467 +0000 UTC m=+144.896631162" watchObservedRunningTime="2025-12-02 13:06:53.944060366 +0000 UTC m=+144.900702061" Dec 02 13:06:53 crc kubenswrapper[4725]: I1202 13:06:53.991140 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:53 crc kubenswrapper[4725]: E1202 13:06:53.992332 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.492311755 +0000 UTC m=+145.448953450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.018015 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" podStartSLOduration=126.017997428 podStartE2EDuration="2m6.017997428s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:54.01644576 +0000 UTC m=+144.973087455" watchObservedRunningTime="2025-12-02 13:06:54.017997428 +0000 UTC m=+144.974639113" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.069284 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2kwpr" podStartSLOduration=10.06926218 podStartE2EDuration="10.06926218s" podCreationTimestamp="2025-12-02 13:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:54.056098131 +0000 UTC m=+145.012739826" watchObservedRunningTime="2025-12-02 13:06:54.06926218 +0000 UTC m=+145.025903875" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.073516 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:54 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:54 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:54 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.073579 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.093322 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.095548 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.595534337 +0000 UTC m=+145.552176102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.172304 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-76gpw" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.202091 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.202364 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.702338915 +0000 UTC m=+145.658980610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.202583 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.202900 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.702892699 +0000 UTC m=+145.659534484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.285960 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.286020 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.304576 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.304969 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.804941642 +0000 UTC m=+145.761583337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.305239 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.305507 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.805498736 +0000 UTC m=+145.762140431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.392977 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w4rh7"] Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.396662 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntm7v"] Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.406422 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.409964 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:54.909940737 +0000 UTC m=+145.866582432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.469771 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lz9d"] Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.514338 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.514675 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.014664195 +0000 UTC m=+145.971305890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.616981 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.617322 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.117300053 +0000 UTC m=+146.073941748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.699602 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brw47"] Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.718941 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.719290 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.219278754 +0000 UTC m=+146.175920449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.820542 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.320521558 +0000 UTC m=+146.277163253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.820578 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.820737 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.821122 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.321114362 +0000 UTC m=+146.277756057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.921837 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.922010 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.421990667 +0000 UTC m=+146.378632362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.922217 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:54 crc kubenswrapper[4725]: E1202 13:06:54.922550 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.42253541 +0000 UTC m=+146.379177115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.925922 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5gxkr"] Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.926993 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.929842 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.947750 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" event={"ID":"bb2e34c8-7e96-44c9-a073-87101442f86e","Type":"ContainerStarted","Data":"0ae8bd0530f0b1be0880cffeb89c93e50b1044e90cefc2b975012c1f5c9b297d"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.947811 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5gxkr"] Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.947834 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" event={"ID":"bb2e34c8-7e96-44c9-a073-87101442f86e","Type":"ContainerStarted","Data":"10ee3ed2737148abe69170090a45bd658e698ba27f207184c1557255dd910531"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.949869 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4jfjh" event={"ID":"96b0990e-4363-4b28-88fa-dfde258940ac","Type":"ContainerStarted","Data":"296744c0e51f7679a3256b903532be972e2427c06a454553cc6248cc86d6ac66"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.949994 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-4jfjh" Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.951143 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" event={"ID":"1de2b4c2-30cb-43ea-bf2a-891bea6af79f","Type":"ContainerStarted","Data":"56ff80baedaf5dacae978114e3d8d6f2dd64d0eba4a3428176131e6648e27fb4"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.952436 4725 generic.go:334] "Generic (PLEG): container finished" podID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerID="06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027" exitCode=0 Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.952518 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerDied","Data":"06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.952534 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerStarted","Data":"664cd16e771996c495d4ec7ba5d4f78b46f545c51c2cd6bc651a6c5a76353925"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.956021 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.966876 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fwlnc" event={"ID":"18606d46-7784-43be-826b-1b1266255201","Type":"ContainerStarted","Data":"2ec11e78214099ad855d66f2969b18d2e6d5bd9704cabe68b193f5cbf972eec9"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.968020 4725 generic.go:334] "Generic (PLEG): container finished" podID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerID="9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc" exitCode=0 Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.968058 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerDied","Data":"9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.968072 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerStarted","Data":"d4d3bb8be5b6117d1a9dd3ad79b39cd04e1d9d1fa6b1362f2a89b3560b49ca0a"} Dec 02 13:06:54 crc kubenswrapper[4725]: I1202 13:06:54.984450 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brw47" event={"ID":"7044a434-25d2-4aed-bcd1-17864366d235","Type":"ContainerStarted","Data":"f05a956eefb98470dff2d3ca9ecc7d8a45c4fae1fb25a3f920f4f27d741cd104"} Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.008426 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" event={"ID":"f3877ac3-8496-4b4f-9578-8ee270d55db2","Type":"ContainerStarted","Data":"a3ac46bd0965367bffc85b4291b5ac1370f1552fb9e1d4f47fff5f28bdd96f57"} Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.014235 4725 generic.go:334] "Generic (PLEG): container finished" podID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerID="0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91" exitCode=0 Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.014338 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntm7v" event={"ID":"362e7a06-0d49-46c4-b7e9-1bfc52125374","Type":"ContainerDied","Data":"0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91"} Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.014365 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntm7v" event={"ID":"362e7a06-0d49-46c4-b7e9-1bfc52125374","Type":"ContainerStarted","Data":"ca9d7dfb9dd0213f43ac64b52f396caf8f1297388dd7b14ef7cd208d2c403c9e"} Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.025501 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" podStartSLOduration=127.025475365 podStartE2EDuration="2m7.025475365s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:55.01702793 +0000 UTC m=+145.973669635" watchObservedRunningTime="2025-12-02 13:06:55.025475365 +0000 UTC m=+145.982117060" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.027783 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.028253 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.528240282 +0000 UTC m=+146.484881977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.035842 4725 generic.go:334] "Generic (PLEG): container finished" podID="66462d0f-425e-470b-a4e2-94682d534fe9" containerID="3a0e3a76fff0d396e51bf96e1542597b76841b39cadaefda8908c7044594e915" exitCode=0 Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.037133 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" event={"ID":"66462d0f-425e-470b-a4e2-94682d534fe9","Type":"ContainerDied","Data":"3a0e3a76fff0d396e51bf96e1542597b76841b39cadaefda8908c7044594e915"} Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.041568 4725 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ldhmm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.041631 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.048989 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:55 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:55 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:55 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.049042 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.055274 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-4jfjh" podStartSLOduration=11.055251407 podStartE2EDuration="11.055251407s" podCreationTimestamp="2025-12-02 13:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:55.050604714 +0000 UTC m=+146.007246409" watchObservedRunningTime="2025-12-02 13:06:55.055251407 +0000 UTC m=+146.011893102" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.057085 4725 patch_prober.go:28] interesting pod/downloads-7954f5f757-gp8pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.057172 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gp8pw" podUID="a81f9c3a-432a-4454-95fa-aea3a8cc3344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.130189 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-catalog-content\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.130468 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqh9b\" (UniqueName: \"kubernetes.io/projected/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-kube-api-access-nqh9b\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.130838 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-utilities\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.131146 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.141142 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.641128458 +0000 UTC m=+146.597770153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.143855 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" podStartSLOduration=127.143837224 podStartE2EDuration="2m7.143837224s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:55.103816664 +0000 UTC m=+146.060458359" watchObservedRunningTime="2025-12-02 13:06:55.143837224 +0000 UTC m=+146.100478919" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.232099 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.232332 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.732302888 +0000 UTC m=+146.688944583 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.232440 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-catalog-content\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.232494 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqh9b\" (UniqueName: \"kubernetes.io/projected/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-kube-api-access-nqh9b\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.232556 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-utilities\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.232600 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.233062 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-utilities\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.233116 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-catalog-content\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.233815 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.733786984 +0000 UTC m=+146.690428759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.245783 4725 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.253311 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqh9b\" (UniqueName: \"kubernetes.io/projected/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-kube-api-access-nqh9b\") pod \"redhat-marketplace-5gxkr\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.298615 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.332420 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rsd6n"] Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.333557 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.333573 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.333916 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.83390084 +0000 UTC m=+146.790542535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.354215 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rsd6n"] Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.434766 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.435097 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-utilities\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.435164 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-catalog-content\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.435217 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wmlx\" (UniqueName: \"kubernetes.io/projected/c90b63b9-848e-458e-9c0e-47474cf7aa10-kube-api-access-4wmlx\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.435239 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 13:06:55.935224226 +0000 UTC m=+146.891865921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5m8fq" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.473993 4725 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-02T13:06:55.245814005Z","Handler":null,"Name":""} Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.536068 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.536554 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wmlx\" (UniqueName: \"kubernetes.io/projected/c90b63b9-848e-458e-9c0e-47474cf7aa10-kube-api-access-4wmlx\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.536641 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-utilities\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.536689 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-catalog-content\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.537060 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-catalog-content\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: E1202 13:06:55.537141 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 13:06:56.037111375 +0000 UTC m=+146.993753070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.537660 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-utilities\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.569661 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wmlx\" (UniqueName: \"kubernetes.io/projected/c90b63b9-848e-458e-9c0e-47474cf7aa10-kube-api-access-4wmlx\") pod \"redhat-marketplace-rsd6n\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.587110 4725 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.587151 4725 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.605034 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5gxkr"] Dec 02 13:06:55 crc kubenswrapper[4725]: W1202 13:06:55.624790 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9517f12e_2c7b_4c89_b65d_a9412e82e9a1.slice/crio-f870921e127edbb57cf2ccd1023032e9988196dbd06fb773b65825cfde7d9045 WatchSource:0}: Error finding container f870921e127edbb57cf2ccd1023032e9988196dbd06fb773b65825cfde7d9045: Status 404 returned error can't find the container with id f870921e127edbb57cf2ccd1023032e9988196dbd06fb773b65825cfde7d9045 Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.638975 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.641948 4725 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.641977 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.661745 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5m8fq\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.690201 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.704378 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.705273 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.710794 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.723522 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.737134 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.740595 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.784333 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.842783 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b386116-e72b-447d-9a6e-494b8b59edc3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.843085 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b386116-e72b-447d-9a6e-494b8b59edc3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.946745 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b386116-e72b-447d-9a6e-494b8b59edc3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.946918 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b386116-e72b-447d-9a6e-494b8b59edc3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.947172 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b386116-e72b-447d-9a6e-494b8b59edc3-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.955860 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.971358 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b386116-e72b-447d-9a6e-494b8b59edc3-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:55 crc kubenswrapper[4725]: I1202 13:06:55.986512 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rsd6n"] Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.047255 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" event={"ID":"1de2b4c2-30cb-43ea-bf2a-891bea6af79f","Type":"ContainerStarted","Data":"c83302b9f83d7039428777d631d6dd2980827fec643e817196b2b9cb302b71c9"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.047301 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" event={"ID":"1de2b4c2-30cb-43ea-bf2a-891bea6af79f","Type":"ContainerStarted","Data":"224ef10208dacdfe102806c3dd1d3656120223395fb60b472e267cb34a000f34"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.047315 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" event={"ID":"1de2b4c2-30cb-43ea-bf2a-891bea6af79f","Type":"ContainerStarted","Data":"ee6300db0800f408d0dbc95880d26be580c6127520be6d0feef5d52a0a60a8a7"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.048990 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:56 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:56 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:56 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.049031 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.054264 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerStarted","Data":"1d206fbdf0dec272ef5b319472f9a80cae8c017f2e3cf29ce500760af9f64465"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.057868 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.058720 4725 generic.go:334] "Generic (PLEG): container finished" podID="7044a434-25d2-4aed-bcd1-17864366d235" containerID="84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d" exitCode=0 Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.059089 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brw47" event={"ID":"7044a434-25d2-4aed-bcd1-17864366d235","Type":"ContainerDied","Data":"84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.060757 4725 generic.go:334] "Generic (PLEG): container finished" podID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerID="0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c" exitCode=0 Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.061437 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5gxkr" event={"ID":"9517f12e-2c7b-4c89-b65d-a9412e82e9a1","Type":"ContainerDied","Data":"0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.061630 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5gxkr" event={"ID":"9517f12e-2c7b-4c89-b65d-a9412e82e9a1","Type":"ContainerStarted","Data":"f870921e127edbb57cf2ccd1023032e9988196dbd06fb773b65825cfde7d9045"} Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.065446 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.072235 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-g49dl" podStartSLOduration=12.072208833 podStartE2EDuration="12.072208833s" podCreationTimestamp="2025-12-02 13:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:56.067428207 +0000 UTC m=+147.024069912" watchObservedRunningTime="2025-12-02 13:06:56.072208833 +0000 UTC m=+147.028850548" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.255999 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.256055 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.256081 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.256127 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.260410 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.263538 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.264037 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.264946 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.281211 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.287922 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.326675 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-42pds"] Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.327844 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.334636 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.352431 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42pds"] Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.415562 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.467132 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-catalog-content\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.467224 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-utilities\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.467248 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xkqg\" (UniqueName: \"kubernetes.io/projected/c3c613fb-ab24-4568-a644-33d9e06063c9-kube-api-access-5xkqg\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.484886 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5m8fq"] Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.486442 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.578948 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66462d0f-425e-470b-a4e2-94682d534fe9-secret-volume\") pod \"66462d0f-425e-470b-a4e2-94682d534fe9\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.579054 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxtxl\" (UniqueName: \"kubernetes.io/projected/66462d0f-425e-470b-a4e2-94682d534fe9-kube-api-access-hxtxl\") pod \"66462d0f-425e-470b-a4e2-94682d534fe9\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.579136 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66462d0f-425e-470b-a4e2-94682d534fe9-config-volume\") pod \"66462d0f-425e-470b-a4e2-94682d534fe9\" (UID: \"66462d0f-425e-470b-a4e2-94682d534fe9\") " Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.579418 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-utilities\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.579451 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xkqg\" (UniqueName: \"kubernetes.io/projected/c3c613fb-ab24-4568-a644-33d9e06063c9-kube-api-access-5xkqg\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.579527 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-catalog-content\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.581248 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66462d0f-425e-470b-a4e2-94682d534fe9-config-volume" (OuterVolumeSpecName: "config-volume") pod "66462d0f-425e-470b-a4e2-94682d534fe9" (UID: "66462d0f-425e-470b-a4e2-94682d534fe9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.582095 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-utilities\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.582580 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-catalog-content\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.603513 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66462d0f-425e-470b-a4e2-94682d534fe9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "66462d0f-425e-470b-a4e2-94682d534fe9" (UID: "66462d0f-425e-470b-a4e2-94682d534fe9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.604626 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66462d0f-425e-470b-a4e2-94682d534fe9-kube-api-access-hxtxl" (OuterVolumeSpecName: "kube-api-access-hxtxl") pod "66462d0f-425e-470b-a4e2-94682d534fe9" (UID: "66462d0f-425e-470b-a4e2-94682d534fe9"). InnerVolumeSpecName "kube-api-access-hxtxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.617908 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xkqg\" (UniqueName: \"kubernetes.io/projected/c3c613fb-ab24-4568-a644-33d9e06063c9-kube-api-access-5xkqg\") pod \"redhat-operators-42pds\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.668372 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.681410 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66462d0f-425e-470b-a4e2-94682d534fe9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.681440 4725 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/66462d0f-425e-470b-a4e2-94682d534fe9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.681454 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxtxl\" (UniqueName: \"kubernetes.io/projected/66462d0f-425e-470b-a4e2-94682d534fe9-kube-api-access-hxtxl\") on node \"crc\" DevicePath \"\"" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.729909 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kdk5h"] Dec 02 13:06:56 crc kubenswrapper[4725]: E1202 13:06:56.744207 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66462d0f-425e-470b-a4e2-94682d534fe9" containerName="collect-profiles" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.744342 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="66462d0f-425e-470b-a4e2-94682d534fe9" containerName="collect-profiles" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.747098 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="66462d0f-425e-470b-a4e2-94682d534fe9" containerName="collect-profiles" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.747783 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kdk5h"] Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.747867 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.749602 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.883121 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcdsz\" (UniqueName: \"kubernetes.io/projected/d0a66991-5ac7-41fd-b9b0-519c239110f6-kube-api-access-wcdsz\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.884066 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-catalog-content\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.884096 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-utilities\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:56 crc kubenswrapper[4725]: W1202 13:06:56.955976 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-c6afdb8827e071881dd2b0616efb59ab116fa28113f0613f2753bf343e510e13 WatchSource:0}: Error finding container c6afdb8827e071881dd2b0616efb59ab116fa28113f0613f2753bf343e510e13: Status 404 returned error can't find the container with id c6afdb8827e071881dd2b0616efb59ab116fa28113f0613f2753bf343e510e13 Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.986073 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcdsz\" (UniqueName: \"kubernetes.io/projected/d0a66991-5ac7-41fd-b9b0-519c239110f6-kube-api-access-wcdsz\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.986123 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-catalog-content\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:56 crc kubenswrapper[4725]: I1202 13:06:56.986155 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-utilities\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:56.996603 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-catalog-content\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:56.996761 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-utilities\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.021816 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcdsz\" (UniqueName: \"kubernetes.io/projected/d0a66991-5ac7-41fd-b9b0-519c239110f6-kube-api-access-wcdsz\") pod \"redhat-operators-kdk5h\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.059807 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:57 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:57 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:57 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.059868 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.067424 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42pds"] Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.080687 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.135275 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b386116-e72b-447d-9a6e-494b8b59edc3","Type":"ContainerStarted","Data":"0659342e0b9ba538d990dc9541291e2462f644a8f1fed4530fa6173459a97c3e"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.148156 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.148196 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx" event={"ID":"66462d0f-425e-470b-a4e2-94682d534fe9","Type":"ContainerDied","Data":"534d3005718794516154251dc7af2d3162fd58c1f9c1ce9efba52faef1e12367"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.148235 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="534d3005718794516154251dc7af2d3162fd58c1f9c1ce9efba52faef1e12367" Dec 02 13:06:57 crc kubenswrapper[4725]: W1202 13:06:57.202031 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-83dcdc8a5bca61382704cc87c0b6a50360588175c731378c7e9a50fb959a9bf4 WatchSource:0}: Error finding container 83dcdc8a5bca61382704cc87c0b6a50360588175c731378c7e9a50fb959a9bf4: Status 404 returned error can't find the container with id 83dcdc8a5bca61382704cc87c0b6a50360588175c731378c7e9a50fb959a9bf4 Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.202298 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" event={"ID":"7828c47b-d717-48f4-ba0a-375358ec3d79","Type":"ContainerStarted","Data":"78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.202349 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" event={"ID":"7828c47b-d717-48f4-ba0a-375358ec3d79","Type":"ContainerStarted","Data":"fe5338e8df5c062e42ccff03660645be848f6dcaaa5a0e594c2c318435bc0229"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.203939 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c6afdb8827e071881dd2b0616efb59ab116fa28113f0613f2753bf343e510e13"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.204565 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.204934 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.208358 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a9f245b902f776c84566dc6f676d9f9463a2e5bb4b5d5c68e575c1bc7d45cfc1"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.213006 4725 generic.go:334] "Generic (PLEG): container finished" podID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerID="a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e" exitCode=0 Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.214343 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerDied","Data":"a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e"} Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.221729 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.283242 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.415573 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kdk5h"] Dec 02 13:06:57 crc kubenswrapper[4725]: W1202 13:06:57.441292 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0a66991_5ac7_41fd_b9b0_519c239110f6.slice/crio-e74ea62fcf00350a0ba2aeec687e9d895567303c1e916e24247482db084adff6 WatchSource:0}: Error finding container e74ea62fcf00350a0ba2aeec687e9d895567303c1e916e24247482db084adff6: Status 404 returned error can't find the container with id e74ea62fcf00350a0ba2aeec687e9d895567303c1e916e24247482db084adff6 Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.483081 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.485905 4725 patch_prober.go:28] interesting pod/console-f9d7485db-7z7tx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.485995 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7z7tx" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.488530 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.753764 4725 patch_prober.go:28] interesting pod/downloads-7954f5f757-gp8pw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.753808 4725 patch_prober.go:28] interesting pod/downloads-7954f5f757-gp8pw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.754168 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gp8pw" podUID="a81f9c3a-432a-4454-95fa-aea3a8cc3344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.754092 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-gp8pw" podUID="a81f9c3a-432a-4454-95fa-aea3a8cc3344" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.845179 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.845245 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:57 crc kubenswrapper[4725]: I1202 13:06:57.865818 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.045691 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.052813 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:58 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:58 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:58 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.052869 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.225125 4725 generic.go:334] "Generic (PLEG): container finished" podID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerID="6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a" exitCode=0 Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.225195 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42pds" event={"ID":"c3c613fb-ab24-4568-a644-33d9e06063c9","Type":"ContainerDied","Data":"6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.225256 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42pds" event={"ID":"c3c613fb-ab24-4568-a644-33d9e06063c9","Type":"ContainerStarted","Data":"b37fe8bed41cb0df89d5e42e0d3260c97f0d676c810d97ab70e56edd9957dd69"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.231712 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fd87231b86c5863651579c5e133f086ae34cb19d1022e74db3286ce84cb80398"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.231742 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"83dcdc8a5bca61382704cc87c0b6a50360588175c731378c7e9a50fb959a9bf4"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.235242 4725 generic.go:334] "Generic (PLEG): container finished" podID="9b386116-e72b-447d-9a6e-494b8b59edc3" containerID="3879b64db87edb546cb7aa2bc66410c8f4da88eb402b2b1895f6a8be45c66157" exitCode=0 Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.235359 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b386116-e72b-447d-9a6e-494b8b59edc3","Type":"ContainerDied","Data":"3879b64db87edb546cb7aa2bc66410c8f4da88eb402b2b1895f6a8be45c66157"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.239145 4725 generic.go:334] "Generic (PLEG): container finished" podID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerID="268e403b866847aca6adef44edef517258d66aecfbd2f10445df64ce7af20563" exitCode=0 Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.239231 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdk5h" event={"ID":"d0a66991-5ac7-41fd-b9b0-519c239110f6","Type":"ContainerDied","Data":"268e403b866847aca6adef44edef517258d66aecfbd2f10445df64ce7af20563"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.239257 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdk5h" event={"ID":"d0a66991-5ac7-41fd-b9b0-519c239110f6","Type":"ContainerStarted","Data":"e74ea62fcf00350a0ba2aeec687e9d895567303c1e916e24247482db084adff6"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.244305 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"58544b00749ec76ca02a62a20ed96e7e3662c12cbf292e28ba0dd06600b262a1"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.245018 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.249643 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"046adace492939bf107d70544d4b0ed714d4dee3ea1686ed32745268814b2499"} Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.255492 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6fvsl" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.264855 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-x7wfx" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.352523 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" podStartSLOduration=130.352503439 podStartE2EDuration="2m10.352503439s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:06:58.347930108 +0000 UTC m=+149.304571813" watchObservedRunningTime="2025-12-02 13:06:58.352503439 +0000 UTC m=+149.309145134" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.906446 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.908998 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.911328 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.922374 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 02 13:06:58 crc kubenswrapper[4725]: I1202 13:06:58.925627 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.044492 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.044844 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.049507 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:06:59 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:06:59 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:06:59 crc kubenswrapper[4725]: healthz check failed Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.049555 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.146364 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.146419 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.146535 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.164982 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.247823 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.260096 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.645189 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.656234 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.760311 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b386116-e72b-447d-9a6e-494b8b59edc3-kubelet-dir\") pod \"9b386116-e72b-447d-9a6e-494b8b59edc3\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.760420 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b386116-e72b-447d-9a6e-494b8b59edc3-kube-api-access\") pod \"9b386116-e72b-447d-9a6e-494b8b59edc3\" (UID: \"9b386116-e72b-447d-9a6e-494b8b59edc3\") " Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.760469 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b386116-e72b-447d-9a6e-494b8b59edc3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9b386116-e72b-447d-9a6e-494b8b59edc3" (UID: "9b386116-e72b-447d-9a6e-494b8b59edc3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.760667 4725 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b386116-e72b-447d-9a6e-494b8b59edc3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.764500 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b386116-e72b-447d-9a6e-494b8b59edc3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9b386116-e72b-447d-9a6e-494b8b59edc3" (UID: "9b386116-e72b-447d-9a6e-494b8b59edc3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:06:59 crc kubenswrapper[4725]: I1202 13:06:59.861955 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b386116-e72b-447d-9a6e-494b8b59edc3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:00 crc kubenswrapper[4725]: I1202 13:07:00.051226 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:00 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:00 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:00 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:00 crc kubenswrapper[4725]: I1202 13:07:00.051287 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:00 crc kubenswrapper[4725]: I1202 13:07:00.266397 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b","Type":"ContainerStarted","Data":"8af2d78629b7a0fb0d396ae4c3bcd76f61681f9de180e6840442808a01bcedae"} Dec 02 13:07:00 crc kubenswrapper[4725]: I1202 13:07:00.268330 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b386116-e72b-447d-9a6e-494b8b59edc3","Type":"ContainerDied","Data":"0659342e0b9ba538d990dc9541291e2462f644a8f1fed4530fa6173459a97c3e"} Dec 02 13:07:00 crc kubenswrapper[4725]: I1202 13:07:00.268384 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0659342e0b9ba538d990dc9541291e2462f644a8f1fed4530fa6173459a97c3e" Dec 02 13:07:00 crc kubenswrapper[4725]: I1202 13:07:00.268901 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 13:07:01 crc kubenswrapper[4725]: I1202 13:07:01.050041 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:01 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:01 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:01 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:01 crc kubenswrapper[4725]: I1202 13:07:01.050103 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:01 crc kubenswrapper[4725]: I1202 13:07:01.294679 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b","Type":"ContainerStarted","Data":"b811bdda22782e1415eabd0e5d9514d81edb38220ca4c4d6fa991744975f6891"} Dec 02 13:07:02 crc kubenswrapper[4725]: I1202 13:07:02.048923 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:02 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:02 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:02 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:02 crc kubenswrapper[4725]: I1202 13:07:02.048990 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:02 crc kubenswrapper[4725]: I1202 13:07:02.302704 4725 generic.go:334] "Generic (PLEG): container finished" podID="d78ac2fc-3b2e-4f02-a077-1a344fea2b4b" containerID="b811bdda22782e1415eabd0e5d9514d81edb38220ca4c4d6fa991744975f6891" exitCode=0 Dec 02 13:07:02 crc kubenswrapper[4725]: I1202 13:07:02.302800 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b","Type":"ContainerDied","Data":"b811bdda22782e1415eabd0e5d9514d81edb38220ca4c4d6fa991744975f6891"} Dec 02 13:07:03 crc kubenswrapper[4725]: I1202 13:07:03.048310 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:03 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:03 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:03 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:03 crc kubenswrapper[4725]: I1202 13:07:03.048556 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:03 crc kubenswrapper[4725]: I1202 13:07:03.060798 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4jfjh" Dec 02 13:07:04 crc kubenswrapper[4725]: I1202 13:07:04.048800 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:04 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:04 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:04 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:04 crc kubenswrapper[4725]: I1202 13:07:04.048871 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:05 crc kubenswrapper[4725]: I1202 13:07:05.048967 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:05 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:05 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:05 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:05 crc kubenswrapper[4725]: I1202 13:07:05.049368 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:06 crc kubenswrapper[4725]: I1202 13:07:06.048751 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:06 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:06 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:06 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:06 crc kubenswrapper[4725]: I1202 13:07:06.048832 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.047779 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:07 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:07 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:07 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.047842 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.475939 4725 patch_prober.go:28] interesting pod/console-f9d7485db-7z7tx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.476428 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7z7tx" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.761001 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-gp8pw" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.850500 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.988496 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kubelet-dir\") pod \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.988587 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d78ac2fc-3b2e-4f02-a077-1a344fea2b4b" (UID: "d78ac2fc-3b2e-4f02-a077-1a344fea2b4b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.988612 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kube-api-access\") pod \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\" (UID: \"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b\") " Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.988954 4725 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:07 crc kubenswrapper[4725]: I1202 13:07:07.994706 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d78ac2fc-3b2e-4f02-a077-1a344fea2b4b" (UID: "d78ac2fc-3b2e-4f02-a077-1a344fea2b4b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:08 crc kubenswrapper[4725]: I1202 13:07:08.049882 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:08 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:08 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:08 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:08 crc kubenswrapper[4725]: I1202 13:07:08.049970 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:08 crc kubenswrapper[4725]: I1202 13:07:08.090641 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d78ac2fc-3b2e-4f02-a077-1a344fea2b4b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:08 crc kubenswrapper[4725]: I1202 13:07:08.359584 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d78ac2fc-3b2e-4f02-a077-1a344fea2b4b","Type":"ContainerDied","Data":"8af2d78629b7a0fb0d396ae4c3bcd76f61681f9de180e6840442808a01bcedae"} Dec 02 13:07:08 crc kubenswrapper[4725]: I1202 13:07:08.359894 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8af2d78629b7a0fb0d396ae4c3bcd76f61681f9de180e6840442808a01bcedae" Dec 02 13:07:08 crc kubenswrapper[4725]: I1202 13:07:08.359632 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 13:07:09 crc kubenswrapper[4725]: I1202 13:07:09.048818 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:09 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:09 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:09 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:09 crc kubenswrapper[4725]: I1202 13:07:09.048879 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:10 crc kubenswrapper[4725]: I1202 13:07:10.048038 4725 patch_prober.go:28] interesting pod/router-default-5444994796-zdq97 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 13:07:10 crc kubenswrapper[4725]: [-]has-synced failed: reason withheld Dec 02 13:07:10 crc kubenswrapper[4725]: [+]process-running ok Dec 02 13:07:10 crc kubenswrapper[4725]: healthz check failed Dec 02 13:07:10 crc kubenswrapper[4725]: I1202 13:07:10.048109 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zdq97" podUID="7a128839-ee0f-4a13-b04d-422d4c76f9ee" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 13:07:10 crc kubenswrapper[4725]: I1202 13:07:10.421990 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:07:10 crc kubenswrapper[4725]: I1202 13:07:10.433178 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9feadd7b-f199-41c3-b908-43774e1cd2f3-metrics-certs\") pod \"network-metrics-daemon-vs5fz\" (UID: \"9feadd7b-f199-41c3-b908-43774e1cd2f3\") " pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:07:10 crc kubenswrapper[4725]: I1202 13:07:10.692584 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vs5fz" Dec 02 13:07:11 crc kubenswrapper[4725]: I1202 13:07:11.049049 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:07:11 crc kubenswrapper[4725]: I1202 13:07:11.051951 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zdq97" Dec 02 13:07:15 crc kubenswrapper[4725]: I1202 13:07:15.962913 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:07:17 crc kubenswrapper[4725]: I1202 13:07:17.494700 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:07:17 crc kubenswrapper[4725]: I1202 13:07:17.498092 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:07:24 crc kubenswrapper[4725]: I1202 13:07:24.285138 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:07:24 crc kubenswrapper[4725]: I1202 13:07:24.285645 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:07:24 crc kubenswrapper[4725]: E1202 13:07:24.872260 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 02 13:07:24 crc kubenswrapper[4725]: E1202 13:07:24.872493 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqh9b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5gxkr_openshift-marketplace(9517f12e-2c7b-4c89-b65d-a9412e82e9a1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 13:07:24 crc kubenswrapper[4725]: E1202 13:07:24.873844 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5gxkr" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" Dec 02 13:07:24 crc kubenswrapper[4725]: E1202 13:07:24.921253 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 02 13:07:24 crc kubenswrapper[4725]: E1202 13:07:24.921394 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4wmlx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rsd6n_openshift-marketplace(c90b63b9-848e-458e-9c0e-47474cf7aa10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 13:07:24 crc kubenswrapper[4725]: E1202 13:07:24.922935 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rsd6n" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.241290 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5gxkr" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.241315 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rsd6n" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.351869 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.352203 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fl5lr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w4rh7_openshift-marketplace(1da4e04b-0c48-48ed-b6cb-b1f2e563e800): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.353753 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-w4rh7" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.456252 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w4rh7" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.466453 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.466626 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qps46,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9lz9d_openshift-marketplace(3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 13:07:26 crc kubenswrapper[4725]: E1202 13:07:26.467811 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9lz9d" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" Dec 02 13:07:26 crc kubenswrapper[4725]: I1202 13:07:26.669820 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vs5fz"] Dec 02 13:07:26 crc kubenswrapper[4725]: W1202 13:07:26.678316 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9feadd7b_f199_41c3_b908_43774e1cd2f3.slice/crio-1901820ce937659021ed9a55ce898a7266c00c108a53de7a3da978c8f6b87fa8 WatchSource:0}: Error finding container 1901820ce937659021ed9a55ce898a7266c00c108a53de7a3da978c8f6b87fa8: Status 404 returned error can't find the container with id 1901820ce937659021ed9a55ce898a7266c00c108a53de7a3da978c8f6b87fa8 Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.453607 4725 generic.go:334] "Generic (PLEG): container finished" podID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerID="5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a" exitCode=0 Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.453766 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42pds" event={"ID":"c3c613fb-ab24-4568-a644-33d9e06063c9","Type":"ContainerDied","Data":"5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a"} Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.455927 4725 generic.go:334] "Generic (PLEG): container finished" podID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerID="cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0" exitCode=0 Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.455979 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntm7v" event={"ID":"362e7a06-0d49-46c4-b7e9-1bfc52125374","Type":"ContainerDied","Data":"cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0"} Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.466969 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" event={"ID":"9feadd7b-f199-41c3-b908-43774e1cd2f3","Type":"ContainerStarted","Data":"80dd7818a6c813eb5a418bf2ec9130517c12c2a0811b89923cc3bb84d1da8840"} Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.467019 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" event={"ID":"9feadd7b-f199-41c3-b908-43774e1cd2f3","Type":"ContainerStarted","Data":"9ef020751e613f1cd6f19d4a1ac4fab2660a06e6d215ea8626ec625118161e38"} Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.467034 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vs5fz" event={"ID":"9feadd7b-f199-41c3-b908-43774e1cd2f3","Type":"ContainerStarted","Data":"1901820ce937659021ed9a55ce898a7266c00c108a53de7a3da978c8f6b87fa8"} Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.471822 4725 generic.go:334] "Generic (PLEG): container finished" podID="7044a434-25d2-4aed-bcd1-17864366d235" containerID="8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c" exitCode=0 Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.472184 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brw47" event={"ID":"7044a434-25d2-4aed-bcd1-17864366d235","Type":"ContainerDied","Data":"8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c"} Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.477709 4725 generic.go:334] "Generic (PLEG): container finished" podID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerID="f04b5faf4d8e276d8e56857747b2decca4ad0998ee3ca667eeb19afbd97853ae" exitCode=0 Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.478084 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdk5h" event={"ID":"d0a66991-5ac7-41fd-b9b0-519c239110f6","Type":"ContainerDied","Data":"f04b5faf4d8e276d8e56857747b2decca4ad0998ee3ca667eeb19afbd97853ae"} Dec 02 13:07:27 crc kubenswrapper[4725]: E1202 13:07:27.478965 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9lz9d" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" Dec 02 13:07:27 crc kubenswrapper[4725]: I1202 13:07:27.503626 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vs5fz" podStartSLOduration=159.503579564 podStartE2EDuration="2m39.503579564s" podCreationTimestamp="2025-12-02 13:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:07:27.500034318 +0000 UTC m=+178.456676013" watchObservedRunningTime="2025-12-02 13:07:27.503579564 +0000 UTC m=+178.460221269" Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.333927 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pg7x8" Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.496436 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntm7v" event={"ID":"362e7a06-0d49-46c4-b7e9-1bfc52125374","Type":"ContainerStarted","Data":"78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971"} Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.498059 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brw47" event={"ID":"7044a434-25d2-4aed-bcd1-17864366d235","Type":"ContainerStarted","Data":"8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46"} Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.500762 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdk5h" event={"ID":"d0a66991-5ac7-41fd-b9b0-519c239110f6","Type":"ContainerStarted","Data":"783b2bf075262d4343a4a44ef981d563d0419f451ebd3babcf55120b7f7b7597"} Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.545114 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ntm7v" podStartSLOduration=2.487558634 podStartE2EDuration="35.545033344s" podCreationTimestamp="2025-12-02 13:06:53 +0000 UTC" firstStartedPulling="2025-12-02 13:06:55.023737883 +0000 UTC m=+145.980379578" lastFinishedPulling="2025-12-02 13:07:28.081212593 +0000 UTC m=+179.037854288" observedRunningTime="2025-12-02 13:07:28.524714551 +0000 UTC m=+179.481356246" watchObservedRunningTime="2025-12-02 13:07:28.545033344 +0000 UTC m=+179.501675039" Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.547071 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kdk5h" podStartSLOduration=2.634334699 podStartE2EDuration="32.547057063s" podCreationTimestamp="2025-12-02 13:06:56 +0000 UTC" firstStartedPulling="2025-12-02 13:06:58.266254628 +0000 UTC m=+149.222896323" lastFinishedPulling="2025-12-02 13:07:28.178976992 +0000 UTC m=+179.135618687" observedRunningTime="2025-12-02 13:07:28.542899272 +0000 UTC m=+179.499540967" watchObservedRunningTime="2025-12-02 13:07:28.547057063 +0000 UTC m=+179.503698758" Dec 02 13:07:28 crc kubenswrapper[4725]: I1202 13:07:28.561174 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-brw47" podStartSLOduration=3.6995631380000003 podStartE2EDuration="35.561160685s" podCreationTimestamp="2025-12-02 13:06:53 +0000 UTC" firstStartedPulling="2025-12-02 13:06:56.096671646 +0000 UTC m=+147.053313341" lastFinishedPulling="2025-12-02 13:07:27.958269193 +0000 UTC m=+178.914910888" observedRunningTime="2025-12-02 13:07:28.560516469 +0000 UTC m=+179.517158164" watchObservedRunningTime="2025-12-02 13:07:28.561160685 +0000 UTC m=+179.517802380" Dec 02 13:07:29 crc kubenswrapper[4725]: I1202 13:07:29.507555 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42pds" event={"ID":"c3c613fb-ab24-4568-a644-33d9e06063c9","Type":"ContainerStarted","Data":"82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977"} Dec 02 13:07:29 crc kubenswrapper[4725]: I1202 13:07:29.534270 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-42pds" podStartSLOduration=3.395450966 podStartE2EDuration="33.534250999s" podCreationTimestamp="2025-12-02 13:06:56 +0000 UTC" firstStartedPulling="2025-12-02 13:06:58.231695471 +0000 UTC m=+149.188337166" lastFinishedPulling="2025-12-02 13:07:28.370495504 +0000 UTC m=+179.327137199" observedRunningTime="2025-12-02 13:07:29.529950845 +0000 UTC m=+180.486592540" watchObservedRunningTime="2025-12-02 13:07:29.534250999 +0000 UTC m=+180.490892694" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.349023 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wst96"] Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.933076 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 13:07:31 crc kubenswrapper[4725]: E1202 13:07:31.933446 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b386116-e72b-447d-9a6e-494b8b59edc3" containerName="pruner" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.933514 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b386116-e72b-447d-9a6e-494b8b59edc3" containerName="pruner" Dec 02 13:07:31 crc kubenswrapper[4725]: E1202 13:07:31.933540 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d78ac2fc-3b2e-4f02-a077-1a344fea2b4b" containerName="pruner" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.933550 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d78ac2fc-3b2e-4f02-a077-1a344fea2b4b" containerName="pruner" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.933688 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d78ac2fc-3b2e-4f02-a077-1a344fea2b4b" containerName="pruner" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.933711 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b386116-e72b-447d-9a6e-494b8b59edc3" containerName="pruner" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.934296 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.939379 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.939821 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 02 13:07:31 crc kubenswrapper[4725]: I1202 13:07:31.951324 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.027099 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/869b7a59-e206-416e-a991-3ac93742f337-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.027348 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/869b7a59-e206-416e-a991-3ac93742f337-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.129159 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/869b7a59-e206-416e-a991-3ac93742f337-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.129238 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/869b7a59-e206-416e-a991-3ac93742f337-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.129364 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/869b7a59-e206-416e-a991-3ac93742f337-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.147759 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/869b7a59-e206-416e-a991-3ac93742f337-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.289265 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:32 crc kubenswrapper[4725]: I1202 13:07:32.702685 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.484750 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.484825 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.532595 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"869b7a59-e206-416e-a991-3ac93742f337","Type":"ContainerStarted","Data":"fd497393c8ef8a529b9f3864e84d0b3772ba82cf6890b7f3c287531a24015609"} Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.533112 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"869b7a59-e206-416e-a991-3ac93742f337","Type":"ContainerStarted","Data":"78f6d9dfc76709c899a7fa3c6d49c910940571a70d125bd219095d2a01dbb9ba"} Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.566925 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.566889484 podStartE2EDuration="2.566889484s" podCreationTimestamp="2025-12-02 13:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:07:33.549670826 +0000 UTC m=+184.506312521" watchObservedRunningTime="2025-12-02 13:07:33.566889484 +0000 UTC m=+184.523531179" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.582667 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.630227 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.929088 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.929327 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:07:33 crc kubenswrapper[4725]: I1202 13:07:33.969035 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:07:34 crc kubenswrapper[4725]: I1202 13:07:34.538862 4725 generic.go:334] "Generic (PLEG): container finished" podID="869b7a59-e206-416e-a991-3ac93742f337" containerID="fd497393c8ef8a529b9f3864e84d0b3772ba82cf6890b7f3c287531a24015609" exitCode=0 Dec 02 13:07:34 crc kubenswrapper[4725]: I1202 13:07:34.538974 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"869b7a59-e206-416e-a991-3ac93742f337","Type":"ContainerDied","Data":"fd497393c8ef8a529b9f3864e84d0b3772ba82cf6890b7f3c287531a24015609"} Dec 02 13:07:34 crc kubenswrapper[4725]: I1202 13:07:34.578837 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.530888 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brw47"] Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.902597 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.976929 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/869b7a59-e206-416e-a991-3ac93742f337-kube-api-access\") pod \"869b7a59-e206-416e-a991-3ac93742f337\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.976989 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/869b7a59-e206-416e-a991-3ac93742f337-kubelet-dir\") pod \"869b7a59-e206-416e-a991-3ac93742f337\" (UID: \"869b7a59-e206-416e-a991-3ac93742f337\") " Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.977145 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/869b7a59-e206-416e-a991-3ac93742f337-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "869b7a59-e206-416e-a991-3ac93742f337" (UID: "869b7a59-e206-416e-a991-3ac93742f337"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.977213 4725 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/869b7a59-e206-416e-a991-3ac93742f337-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:35 crc kubenswrapper[4725]: I1202 13:07:35.982398 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869b7a59-e206-416e-a991-3ac93742f337-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "869b7a59-e206-416e-a991-3ac93742f337" (UID: "869b7a59-e206-416e-a991-3ac93742f337"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.078940 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/869b7a59-e206-416e-a991-3ac93742f337-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.292550 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.548657 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"869b7a59-e206-416e-a991-3ac93742f337","Type":"ContainerDied","Data":"78f6d9dfc76709c899a7fa3c6d49c910940571a70d125bd219095d2a01dbb9ba"} Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.548707 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f6d9dfc76709c899a7fa3c6d49c910940571a70d125bd219095d2a01dbb9ba" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.548673 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.548822 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-brw47" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="registry-server" containerID="cri-o://8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46" gracePeriod=2 Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.669250 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.669298 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:07:36 crc kubenswrapper[4725]: I1202 13:07:36.706472 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.081357 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.081756 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.125473 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.487719 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.558589 4725 generic.go:334] "Generic (PLEG): container finished" podID="7044a434-25d2-4aed-bcd1-17864366d235" containerID="8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46" exitCode=0 Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.558647 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brw47" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.558672 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brw47" event={"ID":"7044a434-25d2-4aed-bcd1-17864366d235","Type":"ContainerDied","Data":"8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46"} Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.558720 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brw47" event={"ID":"7044a434-25d2-4aed-bcd1-17864366d235","Type":"ContainerDied","Data":"f05a956eefb98470dff2d3ca9ecc7d8a45c4fae1fb25a3f920f4f27d741cd104"} Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.558741 4725 scope.go:117] "RemoveContainer" containerID="8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.575579 4725 scope.go:117] "RemoveContainer" containerID="8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.600058 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-utilities\") pod \"7044a434-25d2-4aed-bcd1-17864366d235\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.600267 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwss2\" (UniqueName: \"kubernetes.io/projected/7044a434-25d2-4aed-bcd1-17864366d235-kube-api-access-nwss2\") pod \"7044a434-25d2-4aed-bcd1-17864366d235\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.600313 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-catalog-content\") pod \"7044a434-25d2-4aed-bcd1-17864366d235\" (UID: \"7044a434-25d2-4aed-bcd1-17864366d235\") " Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.602940 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-utilities" (OuterVolumeSpecName: "utilities") pod "7044a434-25d2-4aed-bcd1-17864366d235" (UID: "7044a434-25d2-4aed-bcd1-17864366d235"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.602963 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.604069 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.610890 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7044a434-25d2-4aed-bcd1-17864366d235-kube-api-access-nwss2" (OuterVolumeSpecName: "kube-api-access-nwss2") pod "7044a434-25d2-4aed-bcd1-17864366d235" (UID: "7044a434-25d2-4aed-bcd1-17864366d235"). InnerVolumeSpecName "kube-api-access-nwss2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.618008 4725 scope.go:117] "RemoveContainer" containerID="84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.655995 4725 scope.go:117] "RemoveContainer" containerID="8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46" Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.656532 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46\": container with ID starting with 8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46 not found: ID does not exist" containerID="8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.656612 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46"} err="failed to get container status \"8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46\": rpc error: code = NotFound desc = could not find container \"8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46\": container with ID starting with 8ed5494a39c593b645f12be9ae2daf47d5a1d1ca6b1d82bea67ef7903a4cff46 not found: ID does not exist" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.656703 4725 scope.go:117] "RemoveContainer" containerID="8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c" Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.657198 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c\": container with ID starting with 8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c not found: ID does not exist" containerID="8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.657234 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c"} err="failed to get container status \"8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c\": rpc error: code = NotFound desc = could not find container \"8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c\": container with ID starting with 8d8d4f0eed02567125015b378806b4af17492f022117c95922ef391e2182921c not found: ID does not exist" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.657263 4725 scope.go:117] "RemoveContainer" containerID="84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d" Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.657589 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d\": container with ID starting with 84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d not found: ID does not exist" containerID="84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.657617 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d"} err="failed to get container status \"84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d\": rpc error: code = NotFound desc = could not find container \"84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d\": container with ID starting with 84f88ae54ea4de922935aa3d8de137049a15fcc86dab8e33ab4d669e624a476d not found: ID does not exist" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.670894 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7044a434-25d2-4aed-bcd1-17864366d235" (UID: "7044a434-25d2-4aed-bcd1-17864366d235"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.702522 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.702565 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7044a434-25d2-4aed-bcd1-17864366d235-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.702578 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwss2\" (UniqueName: \"kubernetes.io/projected/7044a434-25d2-4aed-bcd1-17864366d235-kube-api-access-nwss2\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737207 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.737472 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="extract-content" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737493 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="extract-content" Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.737514 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="extract-utilities" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737524 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="extract-utilities" Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.737535 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869b7a59-e206-416e-a991-3ac93742f337" containerName="pruner" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737542 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="869b7a59-e206-416e-a991-3ac93742f337" containerName="pruner" Dec 02 13:07:37 crc kubenswrapper[4725]: E1202 13:07:37.737559 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="registry-server" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737567 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="registry-server" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737714 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7044a434-25d2-4aed-bcd1-17864366d235" containerName="registry-server" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.737731 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="869b7a59-e206-416e-a991-3ac93742f337" containerName="pruner" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.738205 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.740124 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.743100 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.747332 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.803371 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-kubelet-dir\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.803539 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276b30df-586f-4eba-a848-5b4230667534-kube-api-access\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.803595 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-var-lock\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.881380 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brw47"] Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.883582 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-brw47"] Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.905036 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276b30df-586f-4eba-a848-5b4230667534-kube-api-access\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.905106 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-var-lock\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.905151 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-kubelet-dir\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.905261 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-kubelet-dir\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.905301 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-var-lock\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:37 crc kubenswrapper[4725]: I1202 13:07:37.920104 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276b30df-586f-4eba-a848-5b4230667534-kube-api-access\") pod \"installer-9-crc\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:38 crc kubenswrapper[4725]: I1202 13:07:38.056389 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:07:38 crc kubenswrapper[4725]: I1202 13:07:38.502142 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 13:07:38 crc kubenswrapper[4725]: W1202 13:07:38.511326 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod276b30df_586f_4eba_a848_5b4230667534.slice/crio-671c5f7665d20f77931b93ed7f80c4887d36496f87ab9fe50bd7a5ff58b6f8f0 WatchSource:0}: Error finding container 671c5f7665d20f77931b93ed7f80c4887d36496f87ab9fe50bd7a5ff58b6f8f0: Status 404 returned error can't find the container with id 671c5f7665d20f77931b93ed7f80c4887d36496f87ab9fe50bd7a5ff58b6f8f0 Dec 02 13:07:38 crc kubenswrapper[4725]: I1202 13:07:38.565637 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276b30df-586f-4eba-a848-5b4230667534","Type":"ContainerStarted","Data":"671c5f7665d20f77931b93ed7f80c4887d36496f87ab9fe50bd7a5ff58b6f8f0"} Dec 02 13:07:38 crc kubenswrapper[4725]: I1202 13:07:38.932013 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kdk5h"] Dec 02 13:07:39 crc kubenswrapper[4725]: I1202 13:07:39.275288 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7044a434-25d2-4aed-bcd1-17864366d235" path="/var/lib/kubelet/pods/7044a434-25d2-4aed-bcd1-17864366d235/volumes" Dec 02 13:07:39 crc kubenswrapper[4725]: I1202 13:07:39.572096 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kdk5h" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="registry-server" containerID="cri-o://783b2bf075262d4343a4a44ef981d563d0419f451ebd3babcf55120b7f7b7597" gracePeriod=2 Dec 02 13:07:40 crc kubenswrapper[4725]: I1202 13:07:40.577097 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276b30df-586f-4eba-a848-5b4230667534","Type":"ContainerStarted","Data":"bc699cf2ffd23e0bd0a3d872334c46eaf05d1c5ef431b60205261127631f7a2b"} Dec 02 13:07:40 crc kubenswrapper[4725]: I1202 13:07:40.580055 4725 generic.go:334] "Generic (PLEG): container finished" podID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerID="783b2bf075262d4343a4a44ef981d563d0419f451ebd3babcf55120b7f7b7597" exitCode=0 Dec 02 13:07:40 crc kubenswrapper[4725]: I1202 13:07:40.580175 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdk5h" event={"ID":"d0a66991-5ac7-41fd-b9b0-519c239110f6","Type":"ContainerDied","Data":"783b2bf075262d4343a4a44ef981d563d0419f451ebd3babcf55120b7f7b7597"} Dec 02 13:07:40 crc kubenswrapper[4725]: I1202 13:07:40.594278 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.594256428 podStartE2EDuration="3.594256428s" podCreationTimestamp="2025-12-02 13:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:07:40.591949282 +0000 UTC m=+191.548590977" watchObservedRunningTime="2025-12-02 13:07:40.594256428 +0000 UTC m=+191.550898123" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.121531 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.236105 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-utilities\") pod \"d0a66991-5ac7-41fd-b9b0-519c239110f6\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.236179 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcdsz\" (UniqueName: \"kubernetes.io/projected/d0a66991-5ac7-41fd-b9b0-519c239110f6-kube-api-access-wcdsz\") pod \"d0a66991-5ac7-41fd-b9b0-519c239110f6\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.236236 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-catalog-content\") pod \"d0a66991-5ac7-41fd-b9b0-519c239110f6\" (UID: \"d0a66991-5ac7-41fd-b9b0-519c239110f6\") " Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.236930 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-utilities" (OuterVolumeSpecName: "utilities") pod "d0a66991-5ac7-41fd-b9b0-519c239110f6" (UID: "d0a66991-5ac7-41fd-b9b0-519c239110f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.242473 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a66991-5ac7-41fd-b9b0-519c239110f6-kube-api-access-wcdsz" (OuterVolumeSpecName: "kube-api-access-wcdsz") pod "d0a66991-5ac7-41fd-b9b0-519c239110f6" (UID: "d0a66991-5ac7-41fd-b9b0-519c239110f6"). InnerVolumeSpecName "kube-api-access-wcdsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.337408 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.337438 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcdsz\" (UniqueName: \"kubernetes.io/projected/d0a66991-5ac7-41fd-b9b0-519c239110f6-kube-api-access-wcdsz\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.355609 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0a66991-5ac7-41fd-b9b0-519c239110f6" (UID: "d0a66991-5ac7-41fd-b9b0-519c239110f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.438271 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0a66991-5ac7-41fd-b9b0-519c239110f6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.596617 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kdk5h" event={"ID":"d0a66991-5ac7-41fd-b9b0-519c239110f6","Type":"ContainerDied","Data":"e74ea62fcf00350a0ba2aeec687e9d895567303c1e916e24247482db084adff6"} Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.596670 4725 scope.go:117] "RemoveContainer" containerID="783b2bf075262d4343a4a44ef981d563d0419f451ebd3babcf55120b7f7b7597" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.596810 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kdk5h" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.600964 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerStarted","Data":"45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d"} Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.602675 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerStarted","Data":"dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2"} Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.604452 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerStarted","Data":"930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8"} Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.606298 4725 generic.go:334] "Generic (PLEG): container finished" podID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerID="ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c" exitCode=0 Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.606326 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5gxkr" event={"ID":"9517f12e-2c7b-4c89-b65d-a9412e82e9a1","Type":"ContainerDied","Data":"ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c"} Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.610169 4725 scope.go:117] "RemoveContainer" containerID="f04b5faf4d8e276d8e56857747b2decca4ad0998ee3ca667eeb19afbd97853ae" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.629479 4725 scope.go:117] "RemoveContainer" containerID="268e403b866847aca6adef44edef517258d66aecfbd2f10445df64ce7af20563" Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.673672 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kdk5h"] Dec 02 13:07:43 crc kubenswrapper[4725]: I1202 13:07:43.681338 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kdk5h"] Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.614421 4725 generic.go:334] "Generic (PLEG): container finished" podID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerID="930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8" exitCode=0 Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.614496 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerDied","Data":"930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8"} Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.617439 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5gxkr" event={"ID":"9517f12e-2c7b-4c89-b65d-a9412e82e9a1","Type":"ContainerStarted","Data":"8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995"} Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.622968 4725 generic.go:334] "Generic (PLEG): container finished" podID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerID="45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d" exitCode=0 Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.623031 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerDied","Data":"45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d"} Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.625214 4725 generic.go:334] "Generic (PLEG): container finished" podID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerID="dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2" exitCode=0 Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.625251 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerDied","Data":"dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2"} Dec 02 13:07:44 crc kubenswrapper[4725]: I1202 13:07:44.651404 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5gxkr" podStartSLOduration=2.486147485 podStartE2EDuration="50.651388057s" podCreationTimestamp="2025-12-02 13:06:54 +0000 UTC" firstStartedPulling="2025-12-02 13:06:56.096152733 +0000 UTC m=+147.052794428" lastFinishedPulling="2025-12-02 13:07:44.261393285 +0000 UTC m=+195.218035000" observedRunningTime="2025-12-02 13:07:44.647646746 +0000 UTC m=+195.604288451" watchObservedRunningTime="2025-12-02 13:07:44.651388057 +0000 UTC m=+195.608029752" Dec 02 13:07:45 crc kubenswrapper[4725]: I1202 13:07:45.295663 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" path="/var/lib/kubelet/pods/d0a66991-5ac7-41fd-b9b0-519c239110f6/volumes" Dec 02 13:07:45 crc kubenswrapper[4725]: I1202 13:07:45.299121 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:07:45 crc kubenswrapper[4725]: I1202 13:07:45.299165 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:07:45 crc kubenswrapper[4725]: I1202 13:07:45.632037 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerStarted","Data":"b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964"} Dec 02 13:07:45 crc kubenswrapper[4725]: I1202 13:07:45.660486 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9lz9d" podStartSLOduration=2.403911246 podStartE2EDuration="52.660442672s" podCreationTimestamp="2025-12-02 13:06:53 +0000 UTC" firstStartedPulling="2025-12-02 13:06:54.971594239 +0000 UTC m=+145.928235934" lastFinishedPulling="2025-12-02 13:07:45.228125665 +0000 UTC m=+196.184767360" observedRunningTime="2025-12-02 13:07:45.652419378 +0000 UTC m=+196.609061083" watchObservedRunningTime="2025-12-02 13:07:45.660442672 +0000 UTC m=+196.617084367" Dec 02 13:07:46 crc kubenswrapper[4725]: I1202 13:07:46.340570 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-5gxkr" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="registry-server" probeResult="failure" output=< Dec 02 13:07:46 crc kubenswrapper[4725]: timeout: failed to connect service ":50051" within 1s Dec 02 13:07:46 crc kubenswrapper[4725]: > Dec 02 13:07:46 crc kubenswrapper[4725]: I1202 13:07:46.639587 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerStarted","Data":"11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa"} Dec 02 13:07:46 crc kubenswrapper[4725]: I1202 13:07:46.642010 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerStarted","Data":"efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965"} Dec 02 13:07:46 crc kubenswrapper[4725]: I1202 13:07:46.655341 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w4rh7" podStartSLOduration=3.991464926 podStartE2EDuration="54.655326654s" podCreationTimestamp="2025-12-02 13:06:52 +0000 UTC" firstStartedPulling="2025-12-02 13:06:54.955757975 +0000 UTC m=+145.912399670" lastFinishedPulling="2025-12-02 13:07:45.619619703 +0000 UTC m=+196.576261398" observedRunningTime="2025-12-02 13:07:46.653612172 +0000 UTC m=+197.610253867" watchObservedRunningTime="2025-12-02 13:07:46.655326654 +0000 UTC m=+197.611968349" Dec 02 13:07:46 crc kubenswrapper[4725]: I1202 13:07:46.672821 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rsd6n" podStartSLOduration=3.107788657 podStartE2EDuration="51.672804567s" podCreationTimestamp="2025-12-02 13:06:55 +0000 UTC" firstStartedPulling="2025-12-02 13:06:57.21995115 +0000 UTC m=+148.176592855" lastFinishedPulling="2025-12-02 13:07:45.78496707 +0000 UTC m=+196.741608765" observedRunningTime="2025-12-02 13:07:46.672558451 +0000 UTC m=+197.629200146" watchObservedRunningTime="2025-12-02 13:07:46.672804567 +0000 UTC m=+197.629446262" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.490673 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.492165 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.535110 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.703828 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.704197 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.712383 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:07:53 crc kubenswrapper[4725]: I1202 13:07:53.743679 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.284741 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.284788 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.284827 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.285308 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.285350 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f" gracePeriod=600 Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.677836 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f" exitCode=0 Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.678721 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f"} Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.678745 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"e34e8f201b66e3cbc21d632890c9a1c51e534112ab1f9ab4f4f771f7eb972cdd"} Dec 02 13:07:54 crc kubenswrapper[4725]: I1202 13:07:54.718538 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:07:55 crc kubenswrapper[4725]: I1202 13:07:55.342772 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:07:55 crc kubenswrapper[4725]: I1202 13:07:55.381086 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:07:55 crc kubenswrapper[4725]: I1202 13:07:55.693588 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:07:55 crc kubenswrapper[4725]: I1202 13:07:55.693623 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:07:55 crc kubenswrapper[4725]: I1202 13:07:55.733971 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:07:56 crc kubenswrapper[4725]: I1202 13:07:56.383267 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" podUID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" containerName="oauth-openshift" containerID="cri-o://78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d" gracePeriod=15 Dec 02 13:07:56 crc kubenswrapper[4725]: I1202 13:07:56.736553 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:07:56 crc kubenswrapper[4725]: I1202 13:07:56.931698 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lz9d"] Dec 02 13:07:56 crc kubenswrapper[4725]: I1202 13:07:56.931909 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9lz9d" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="registry-server" containerID="cri-o://b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964" gracePeriod=2 Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.362015 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.367254 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388153 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2"] Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388371 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="registry-server" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388382 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="registry-server" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388392 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="extract-utilities" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388398 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="extract-utilities" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388407 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="extract-utilities" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388413 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="extract-utilities" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388424 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="extract-content" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388429 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="extract-content" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388436 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" containerName="oauth-openshift" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388441 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" containerName="oauth-openshift" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388450 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="registry-server" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388501 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="registry-server" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.388514 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="extract-content" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388520 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="extract-content" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388601 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerName="registry-server" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388613 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" containerName="oauth-openshift" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388625 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a66991-5ac7-41fd-b9b0-519c239110f6" containerName="registry-server" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.388994 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.407388 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2"] Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444783 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-catalog-content\") pod \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444843 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-dir\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444865 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-trusted-ca-bundle\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444885 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-session\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444907 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qps46\" (UniqueName: \"kubernetes.io/projected/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-kube-api-access-qps46\") pod \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444934 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-idp-0-file-data\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444950 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-cliconfig\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.444972 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlhfd\" (UniqueName: \"kubernetes.io/projected/2034b4c1-5d0f-42d4-951d-afd68a259ee1-kube-api-access-xlhfd\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445000 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-router-certs\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445016 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-serving-cert\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445033 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-login\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445060 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-utilities\") pod \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\" (UID: \"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445080 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-ocp-branding-template\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445099 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-provider-selection\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445115 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-policies\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445129 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-error\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445160 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-service-ca\") pod \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\" (UID: \"2034b4c1-5d0f-42d4-951d-afd68a259ee1\") " Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445250 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-error\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445279 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445296 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445318 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-service-ca\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445340 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85a74938-50c8-43b7-afc4-f4eadd5f84b2-audit-dir\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445355 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2tnf\" (UniqueName: \"kubernetes.io/projected/85a74938-50c8-43b7-afc4-f4eadd5f84b2-kube-api-access-s2tnf\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445379 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-router-certs\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445396 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445414 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-session\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445435 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-login\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445483 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445511 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445529 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445550 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-audit-policies\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445812 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445810 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-utilities" (OuterVolumeSpecName: "utilities") pod "3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" (UID: "3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.445871 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.448037 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.448166 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.448847 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.457956 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.461680 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.461960 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.462109 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.465266 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.465635 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2034b4c1-5d0f-42d4-951d-afd68a259ee1-kube-api-access-xlhfd" (OuterVolumeSpecName: "kube-api-access-xlhfd") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "kube-api-access-xlhfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.465810 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-kube-api-access-qps46" (OuterVolumeSpecName: "kube-api-access-qps46") pod "3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" (UID: "3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d"). InnerVolumeSpecName "kube-api-access-qps46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.466037 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.467036 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.467219 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2034b4c1-5d0f-42d4-951d-afd68a259ee1" (UID: "2034b4c1-5d0f-42d4-951d-afd68a259ee1"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.499294 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" (UID: "3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546584 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-service-ca\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546638 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85a74938-50c8-43b7-afc4-f4eadd5f84b2-audit-dir\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546668 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2tnf\" (UniqueName: \"kubernetes.io/projected/85a74938-50c8-43b7-afc4-f4eadd5f84b2-kube-api-access-s2tnf\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546704 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-router-certs\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546724 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/85a74938-50c8-43b7-afc4-f4eadd5f84b2-audit-dir\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546729 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546790 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-session\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546817 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-login\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546861 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546885 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546902 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546928 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-audit-policies\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546952 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-error\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.546985 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547008 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547053 4725 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547067 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547079 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547093 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qps46\" (UniqueName: \"kubernetes.io/projected/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-kube-api-access-qps46\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547106 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547118 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547129 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlhfd\" (UniqueName: \"kubernetes.io/projected/2034b4c1-5d0f-42d4-951d-afd68a259ee1-kube-api-access-xlhfd\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547141 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547152 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547165 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547179 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547191 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547202 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547214 4725 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547225 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547237 4725 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2034b4c1-5d0f-42d4-951d-afd68a259ee1-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547252 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547490 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-service-ca\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.547813 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.549822 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.550619 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.550808 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/85a74938-50c8-43b7-afc4-f4eadd5f84b2-audit-policies\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.551239 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.551928 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-session\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.552209 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.552245 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-router-certs\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.552715 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.553935 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-error\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.563167 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2tnf\" (UniqueName: \"kubernetes.io/projected/85a74938-50c8-43b7-afc4-f4eadd5f84b2-kube-api-access-s2tnf\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.564435 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/85a74938-50c8-43b7-afc4-f4eadd5f84b2-v4-0-config-user-template-login\") pod \"oauth-openshift-68c4c8b9fb-rqcd2\" (UID: \"85a74938-50c8-43b7-afc4-f4eadd5f84b2\") " pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.695830 4725 generic.go:334] "Generic (PLEG): container finished" podID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" containerID="78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d" exitCode=0 Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.695904 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.695928 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" event={"ID":"2034b4c1-5d0f-42d4-951d-afd68a259ee1","Type":"ContainerDied","Data":"78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d"} Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.695979 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wst96" event={"ID":"2034b4c1-5d0f-42d4-951d-afd68a259ee1","Type":"ContainerDied","Data":"a4d2f7e538f296ab79e03846aa897c83a86dbf9b3afc25691eff6aaf2fdbdf10"} Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.696001 4725 scope.go:117] "RemoveContainer" containerID="78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.699937 4725 generic.go:334] "Generic (PLEG): container finished" podID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" containerID="b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964" exitCode=0 Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.700960 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lz9d" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.702619 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerDied","Data":"b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964"} Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.702739 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lz9d" event={"ID":"3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d","Type":"ContainerDied","Data":"d4d3bb8be5b6117d1a9dd3ad79b39cd04e1d9d1fa6b1362f2a89b3560b49ca0a"} Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.713346 4725 scope.go:117] "RemoveContainer" containerID="78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.713831 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d\": container with ID starting with 78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d not found: ID does not exist" containerID="78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.713865 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d"} err="failed to get container status \"78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d\": rpc error: code = NotFound desc = could not find container \"78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d\": container with ID starting with 78e26451fa97af4fb4c169e80376998ee56aa71eb92b687feefb71bfc5d8373d not found: ID does not exist" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.713897 4725 scope.go:117] "RemoveContainer" containerID="b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.727968 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.737487 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wst96"] Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.746870 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wst96"] Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.747404 4725 scope.go:117] "RemoveContainer" containerID="930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.763195 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lz9d"] Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.764427 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9lz9d"] Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.791251 4725 scope.go:117] "RemoveContainer" containerID="9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.806620 4725 scope.go:117] "RemoveContainer" containerID="b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.806952 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964\": container with ID starting with b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964 not found: ID does not exist" containerID="b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.806989 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964"} err="failed to get container status \"b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964\": rpc error: code = NotFound desc = could not find container \"b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964\": container with ID starting with b166c48ef3b50d01f349e26e09707be43ba4c7e6916b94695f24016c6c471964 not found: ID does not exist" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.807016 4725 scope.go:117] "RemoveContainer" containerID="930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.807200 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8\": container with ID starting with 930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8 not found: ID does not exist" containerID="930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.807221 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8"} err="failed to get container status \"930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8\": rpc error: code = NotFound desc = could not find container \"930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8\": container with ID starting with 930fa7260f3a8b15e742de041f726f9eb1fd0eb78c18493719ffd1bb241884b8 not found: ID does not exist" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.807234 4725 scope.go:117] "RemoveContainer" containerID="9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc" Dec 02 13:07:57 crc kubenswrapper[4725]: E1202 13:07:57.807380 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc\": container with ID starting with 9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc not found: ID does not exist" containerID="9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.807400 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc"} err="failed to get container status \"9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc\": rpc error: code = NotFound desc = could not find container \"9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc\": container with ID starting with 9292ac383408bfa1ac97111735254f8d2849523634bc2b7a50c49d6d515b2dbc not found: ID does not exist" Dec 02 13:07:57 crc kubenswrapper[4725]: I1202 13:07:57.930992 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rsd6n"] Dec 02 13:07:58 crc kubenswrapper[4725]: I1202 13:07:58.133575 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2"] Dec 02 13:07:58 crc kubenswrapper[4725]: W1202 13:07:58.141503 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85a74938_50c8_43b7_afc4_f4eadd5f84b2.slice/crio-d85025a99bc18032a1df894a2edb7cceff43efb93f8150aa14b61d8bf27f8a6c WatchSource:0}: Error finding container d85025a99bc18032a1df894a2edb7cceff43efb93f8150aa14b61d8bf27f8a6c: Status 404 returned error can't find the container with id d85025a99bc18032a1df894a2edb7cceff43efb93f8150aa14b61d8bf27f8a6c Dec 02 13:07:58 crc kubenswrapper[4725]: I1202 13:07:58.708779 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" event={"ID":"85a74938-50c8-43b7-afc4-f4eadd5f84b2","Type":"ContainerStarted","Data":"639e3de1c6a2bd974f161e92ae6769443d8ccef299b9b3894933b95a1473a95b"} Dec 02 13:07:58 crc kubenswrapper[4725]: I1202 13:07:58.709403 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" event={"ID":"85a74938-50c8-43b7-afc4-f4eadd5f84b2","Type":"ContainerStarted","Data":"d85025a99bc18032a1df894a2edb7cceff43efb93f8150aa14b61d8bf27f8a6c"} Dec 02 13:07:58 crc kubenswrapper[4725]: I1202 13:07:58.708971 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rsd6n" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="registry-server" containerID="cri-o://efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965" gracePeriod=2 Dec 02 13:07:58 crc kubenswrapper[4725]: I1202 13:07:58.733620 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" podStartSLOduration=27.73360101 podStartE2EDuration="27.73360101s" podCreationTimestamp="2025-12-02 13:07:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:07:58.732864639 +0000 UTC m=+209.689506334" watchObservedRunningTime="2025-12-02 13:07:58.73360101 +0000 UTC m=+209.690242695" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.059219 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.164438 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-catalog-content\") pod \"c90b63b9-848e-458e-9c0e-47474cf7aa10\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.164536 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wmlx\" (UniqueName: \"kubernetes.io/projected/c90b63b9-848e-458e-9c0e-47474cf7aa10-kube-api-access-4wmlx\") pod \"c90b63b9-848e-458e-9c0e-47474cf7aa10\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.164559 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-utilities\") pod \"c90b63b9-848e-458e-9c0e-47474cf7aa10\" (UID: \"c90b63b9-848e-458e-9c0e-47474cf7aa10\") " Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.165330 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-utilities" (OuterVolumeSpecName: "utilities") pod "c90b63b9-848e-458e-9c0e-47474cf7aa10" (UID: "c90b63b9-848e-458e-9c0e-47474cf7aa10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.170751 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c90b63b9-848e-458e-9c0e-47474cf7aa10-kube-api-access-4wmlx" (OuterVolumeSpecName: "kube-api-access-4wmlx") pod "c90b63b9-848e-458e-9c0e-47474cf7aa10" (UID: "c90b63b9-848e-458e-9c0e-47474cf7aa10"). InnerVolumeSpecName "kube-api-access-4wmlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.185759 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c90b63b9-848e-458e-9c0e-47474cf7aa10" (UID: "c90b63b9-848e-458e-9c0e-47474cf7aa10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.266364 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.266415 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wmlx\" (UniqueName: \"kubernetes.io/projected/c90b63b9-848e-458e-9c0e-47474cf7aa10-kube-api-access-4wmlx\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.266428 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c90b63b9-848e-458e-9c0e-47474cf7aa10-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.274479 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2034b4c1-5d0f-42d4-951d-afd68a259ee1" path="/var/lib/kubelet/pods/2034b4c1-5d0f-42d4-951d-afd68a259ee1/volumes" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.275167 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d" path="/var/lib/kubelet/pods/3ac2978d-3b11-4496-b2d9-ed7ca3a5e12d/volumes" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.762982 4725 generic.go:334] "Generic (PLEG): container finished" podID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerID="efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965" exitCode=0 Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.763013 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rsd6n" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.763062 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerDied","Data":"efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965"} Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.763114 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rsd6n" event={"ID":"c90b63b9-848e-458e-9c0e-47474cf7aa10","Type":"ContainerDied","Data":"1d206fbdf0dec272ef5b319472f9a80cae8c017f2e3cf29ce500760af9f64465"} Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.763136 4725 scope.go:117] "RemoveContainer" containerID="efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.763987 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.769292 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68c4c8b9fb-rqcd2" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.782293 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rsd6n"] Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.785324 4725 scope.go:117] "RemoveContainer" containerID="dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.786215 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rsd6n"] Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.803402 4725 scope.go:117] "RemoveContainer" containerID="a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.839181 4725 scope.go:117] "RemoveContainer" containerID="efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965" Dec 02 13:07:59 crc kubenswrapper[4725]: E1202 13:07:59.840147 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965\": container with ID starting with efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965 not found: ID does not exist" containerID="efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.840190 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965"} err="failed to get container status \"efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965\": rpc error: code = NotFound desc = could not find container \"efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965\": container with ID starting with efdf7fc790c7ad1238cdb67e06ce020af3631f861bf1a554a425863540944965 not found: ID does not exist" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.840222 4725 scope.go:117] "RemoveContainer" containerID="dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2" Dec 02 13:07:59 crc kubenswrapper[4725]: E1202 13:07:59.840660 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2\": container with ID starting with dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2 not found: ID does not exist" containerID="dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.840683 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2"} err="failed to get container status \"dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2\": rpc error: code = NotFound desc = could not find container \"dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2\": container with ID starting with dc432de80d14139d37759a8b1528b402e4d3c0a5b88a2b2e337dc4a9f85aace2 not found: ID does not exist" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.840697 4725 scope.go:117] "RemoveContainer" containerID="a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e" Dec 02 13:07:59 crc kubenswrapper[4725]: E1202 13:07:59.841169 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e\": container with ID starting with a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e not found: ID does not exist" containerID="a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e" Dec 02 13:07:59 crc kubenswrapper[4725]: I1202 13:07:59.841264 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e"} err="failed to get container status \"a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e\": rpc error: code = NotFound desc = could not find container \"a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e\": container with ID starting with a27a7cc13135450558bc2ce427b498bdba38bdd81c11ce8b21f152ea70c1de3e not found: ID does not exist" Dec 02 13:08:01 crc kubenswrapper[4725]: I1202 13:08:01.274691 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" path="/var/lib/kubelet/pods/c90b63b9-848e-458e-9c0e-47474cf7aa10/volumes" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.609540 4725 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.610985 4725 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611210 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611227 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611236 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="registry-server" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611242 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="registry-server" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611252 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611259 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611267 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611273 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611283 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="extract-utilities" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611289 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="extract-utilities" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611298 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611304 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611311 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611317 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611326 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="extract-content" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611331 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="extract-content" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611339 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611344 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611433 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="c90b63b9-848e-458e-9c0e-47474cf7aa10" containerName="registry-server" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611443 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611452 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611477 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611486 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611493 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611501 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.611608 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.611617 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.612592 4725 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.613008 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.616648 4725 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.647110 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782501 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782551 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782571 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782589 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782740 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782800 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782942 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.782982 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870021 4725 generic.go:334] "Generic (PLEG): container finished" podID="276b30df-586f-4eba-a848-5b4230667534" containerID="bc699cf2ffd23e0bd0a3d872334c46eaf05d1c5ef431b60205261127631f7a2b" exitCode=0 Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870207 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040" gracePeriod=15 Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870297 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276b30df-586f-4eba-a848-5b4230667534","Type":"ContainerDied","Data":"bc699cf2ffd23e0bd0a3d872334c46eaf05d1c5ef431b60205261127631f7a2b"} Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870629 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091" gracePeriod=15 Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870688 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a" gracePeriod=15 Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870723 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1" gracePeriod=15 Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.870761 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a" gracePeriod=15 Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.885212 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.885262 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.885289 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.885991 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.886294 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.886323 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.886372 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.888892 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890088 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890128 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890152 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890176 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890196 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890216 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890237 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.890255 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: I1202 13:08:17.936220 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:08:17 crc kubenswrapper[4725]: W1202 13:08:17.955560 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-5337f7d876fa9a9a0b08a1dfd87557b8e83ecd9c2fb8fddb0d33e1ad829b7e1e WatchSource:0}: Error finding container 5337f7d876fa9a9a0b08a1dfd87557b8e83ecd9c2fb8fddb0d33e1ad829b7e1e: Status 404 returned error can't find the container with id 5337f7d876fa9a9a0b08a1dfd87557b8e83ecd9c2fb8fddb0d33e1ad829b7e1e Dec 02 13:08:17 crc kubenswrapper[4725]: E1202 13:08:17.960164 4725 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d67ed50c13eb7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 13:08:17.958854327 +0000 UTC m=+228.915496022,LastTimestamp:2025-12-02 13:08:17.958854327 +0000 UTC m=+228.915496022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.875950 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e"} Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.876000 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5337f7d876fa9a9a0b08a1dfd87557b8e83ecd9c2fb8fddb0d33e1ad829b7e1e"} Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.878440 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.879543 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.880255 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091" exitCode=0 Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.880283 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a" exitCode=0 Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.880292 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1" exitCode=0 Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.880302 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a" exitCode=2 Dec 02 13:08:18 crc kubenswrapper[4725]: I1202 13:08:18.880343 4725 scope.go:117] "RemoveContainer" containerID="e06019f09732165e7f86e5348e8fce281c9d437da226e5a89994511d6c0bf9ab" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.103980 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.107853 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-var-lock\") pod \"276b30df-586f-4eba-a848-5b4230667534\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.107962 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-var-lock" (OuterVolumeSpecName: "var-lock") pod "276b30df-586f-4eba-a848-5b4230667534" (UID: "276b30df-586f-4eba-a848-5b4230667534"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.108007 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-kubelet-dir\") pod \"276b30df-586f-4eba-a848-5b4230667534\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.108074 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "276b30df-586f-4eba-a848-5b4230667534" (UID: "276b30df-586f-4eba-a848-5b4230667534"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.108237 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276b30df-586f-4eba-a848-5b4230667534-kube-api-access\") pod \"276b30df-586f-4eba-a848-5b4230667534\" (UID: \"276b30df-586f-4eba-a848-5b4230667534\") " Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.109234 4725 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-var-lock\") on node \"crc\" DevicePath \"\"" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.109285 4725 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/276b30df-586f-4eba-a848-5b4230667534-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.112830 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/276b30df-586f-4eba-a848-5b4230667534-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "276b30df-586f-4eba-a848-5b4230667534" (UID: "276b30df-586f-4eba-a848-5b4230667534"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.209693 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/276b30df-586f-4eba-a848-5b4230667534-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.886683 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"276b30df-586f-4eba-a848-5b4230667534","Type":"ContainerDied","Data":"671c5f7665d20f77931b93ed7f80c4887d36496f87ab9fe50bd7a5ff58b6f8f0"} Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.886701 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.886724 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="671c5f7665d20f77931b93ed7f80c4887d36496f87ab9fe50bd7a5ff58b6f8f0" Dec 02 13:08:19 crc kubenswrapper[4725]: I1202 13:08:19.889554 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.231133 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.232355 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.323991 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324023 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324090 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324128 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324196 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324275 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324379 4725 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324396 4725 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.324408 4725 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:08:20 crc kubenswrapper[4725]: E1202 13:08:20.342497 4725 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.18:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d67ed50c13eb7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 13:08:17.958854327 +0000 UTC m=+228.915496022,LastTimestamp:2025-12-02 13:08:17.958854327 +0000 UTC m=+228.915496022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.900759 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.902503 4725 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040" exitCode=0 Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.902555 4725 scope.go:117] "RemoveContainer" containerID="ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.902604 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.923227 4725 scope.go:117] "RemoveContainer" containerID="253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.938142 4725 scope.go:117] "RemoveContainer" containerID="02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.957597 4725 scope.go:117] "RemoveContainer" containerID="7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.971853 4725 scope.go:117] "RemoveContainer" containerID="ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040" Dec 02 13:08:20 crc kubenswrapper[4725]: I1202 13:08:20.985393 4725 scope.go:117] "RemoveContainer" containerID="2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.007409 4725 scope.go:117] "RemoveContainer" containerID="ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091" Dec 02 13:08:21 crc kubenswrapper[4725]: E1202 13:08:21.008690 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\": container with ID starting with ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091 not found: ID does not exist" containerID="ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.008720 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091"} err="failed to get container status \"ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\": rpc error: code = NotFound desc = could not find container \"ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091\": container with ID starting with ecfdc54a4a173be5c5996f39990121b4984f25255cc2a46882cabcbaac5bc091 not found: ID does not exist" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.008742 4725 scope.go:117] "RemoveContainer" containerID="253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a" Dec 02 13:08:21 crc kubenswrapper[4725]: E1202 13:08:21.009111 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\": container with ID starting with 253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a not found: ID does not exist" containerID="253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.009139 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a"} err="failed to get container status \"253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\": rpc error: code = NotFound desc = could not find container \"253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a\": container with ID starting with 253528645fb5c5658790e04f8a78061603777e51f6ec5233d9b57c5fa5d2e25a not found: ID does not exist" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.009158 4725 scope.go:117] "RemoveContainer" containerID="02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1" Dec 02 13:08:21 crc kubenswrapper[4725]: E1202 13:08:21.009384 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\": container with ID starting with 02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1 not found: ID does not exist" containerID="02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.009419 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1"} err="failed to get container status \"02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\": rpc error: code = NotFound desc = could not find container \"02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1\": container with ID starting with 02e4790b38bdae721f99b640faac2d79a7ed28f5693cf43f569503d53074b0b1 not found: ID does not exist" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.009439 4725 scope.go:117] "RemoveContainer" containerID="7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a" Dec 02 13:08:21 crc kubenswrapper[4725]: E1202 13:08:21.009828 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\": container with ID starting with 7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a not found: ID does not exist" containerID="7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.009883 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a"} err="failed to get container status \"7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\": rpc error: code = NotFound desc = could not find container \"7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a\": container with ID starting with 7f6e59333f983291ac306a258054a54aa936e7e559d51056b6b9bbb3c36b5e6a not found: ID does not exist" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.009919 4725 scope.go:117] "RemoveContainer" containerID="ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040" Dec 02 13:08:21 crc kubenswrapper[4725]: E1202 13:08:21.010167 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\": container with ID starting with ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040 not found: ID does not exist" containerID="ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.010192 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040"} err="failed to get container status \"ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\": rpc error: code = NotFound desc = could not find container \"ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040\": container with ID starting with ae411904cb292983cf98c1c32e1ae3fca8f929765bb8bb86e1e62a2d00a23040 not found: ID does not exist" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.010207 4725 scope.go:117] "RemoveContainer" containerID="2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173" Dec 02 13:08:21 crc kubenswrapper[4725]: E1202 13:08:21.010413 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\": container with ID starting with 2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173 not found: ID does not exist" containerID="2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.010437 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173"} err="failed to get container status \"2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\": rpc error: code = NotFound desc = could not find container \"2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173\": container with ID starting with 2ecb5844e9a97fe3906f1a6064725ba7bd70ed07b8e36eb6ee20dd22b5466173 not found: ID does not exist" Dec 02 13:08:21 crc kubenswrapper[4725]: I1202 13:08:21.277385 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.768222 4725 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.768983 4725 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.769241 4725 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.769523 4725 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.769781 4725 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.769814 4725 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.770069 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="200ms" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.893307 4725 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.893737 4725 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.897371 4725 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.897587 4725 status_manager.go:851] "Failed to get status for pod" podUID="276b30df-586f-4eba-a848-5b4230667534" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.897825 4725 status_manager.go:851] "Failed to get status for pod" podUID="276b30df-586f-4eba-a848-5b4230667534" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: I1202 13:08:22.898023 4725 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:22 crc kubenswrapper[4725]: E1202 13:08:22.971206 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="400ms" Dec 02 13:08:23 crc kubenswrapper[4725]: E1202 13:08:23.372616 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="800ms" Dec 02 13:08:24 crc kubenswrapper[4725]: E1202 13:08:24.174171 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="1.6s" Dec 02 13:08:25 crc kubenswrapper[4725]: E1202 13:08:25.775375 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="3.2s" Dec 02 13:08:28 crc kubenswrapper[4725]: E1202 13:08:28.976200 4725 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.18:6443: connect: connection refused" interval="6.4s" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.271539 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.274651 4725 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.275154 4725 status_manager.go:851] "Failed to get status for pod" podUID="276b30df-586f-4eba-a848-5b4230667534" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.276114 4725 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.276320 4725 status_manager.go:851] "Failed to get status for pod" podUID="276b30df-586f-4eba-a848-5b4230667534" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.295626 4725 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.295774 4725 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:29 crc kubenswrapper[4725]: E1202 13:08:29.296282 4725 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.296737 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:29 crc kubenswrapper[4725]: W1202 13:08:29.324776 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-c89553406a2b35e89951c53ab0e4f2e3c0fabb2616b049dd3cfba427d0d880fc WatchSource:0}: Error finding container c89553406a2b35e89951c53ab0e4f2e3c0fabb2616b049dd3cfba427d0d880fc: Status 404 returned error can't find the container with id c89553406a2b35e89951c53ab0e4f2e3c0fabb2616b049dd3cfba427d0d880fc Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.951336 4725 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="15c27f449ed2ba7997b124d7d701e04884e6e171c8e2e88996a07783ab5eb543" exitCode=0 Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.951415 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"15c27f449ed2ba7997b124d7d701e04884e6e171c8e2e88996a07783ab5eb543"} Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.951880 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c89553406a2b35e89951c53ab0e4f2e3c0fabb2616b049dd3cfba427d0d880fc"} Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.952221 4725 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.952242 4725 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.952764 4725 status_manager.go:851] "Failed to get status for pod" podUID="276b30df-586f-4eba-a848-5b4230667534" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:29 crc kubenswrapper[4725]: E1202 13:08:29.952797 4725 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:29 crc kubenswrapper[4725]: I1202 13:08:29.953151 4725 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.18:6443: connect: connection refused" Dec 02 13:08:30 crc kubenswrapper[4725]: I1202 13:08:30.968222 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"efd3a3d12b19854d71b1a1a2b25d34b2aed541a6ddee964e3b45f02c715d1f8b"} Dec 02 13:08:30 crc kubenswrapper[4725]: I1202 13:08:30.968349 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e09274d8265dd1b3b64ec39037571cf726fd8571da6e8d5b3febc3bcac27d2d"} Dec 02 13:08:30 crc kubenswrapper[4725]: I1202 13:08:30.968365 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"935935a2d65f3acec14467dea6e6822ae60afdabc345e456672649edb647fc55"} Dec 02 13:08:30 crc kubenswrapper[4725]: I1202 13:08:30.968378 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b4fa0e251396359cc91a625df393106d9c3b87011db8c73b983a9e02d700b100"} Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.975263 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.975515 4725 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f" exitCode=1 Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.975546 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f"} Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.976031 4725 scope.go:117] "RemoveContainer" containerID="ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f" Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.978183 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"129a97563f960bd50ff9adc59e0a31d01882d8aa6f7d8b838f341ee7e511b3c3"} Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.978312 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.978372 4725 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:31 crc kubenswrapper[4725]: I1202 13:08:31.978388 4725 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:33 crc kubenswrapper[4725]: I1202 13:08:33.041987 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 13:08:33 crc kubenswrapper[4725]: I1202 13:08:33.042043 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f91b606eb7ca216cf92a16ebc9fd6089d8ffd36dc3c8f5e2daec88673cfec644"} Dec 02 13:08:34 crc kubenswrapper[4725]: I1202 13:08:34.297945 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:34 crc kubenswrapper[4725]: I1202 13:08:34.298865 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:34 crc kubenswrapper[4725]: I1202 13:08:34.303318 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:36 crc kubenswrapper[4725]: I1202 13:08:36.987228 4725 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:37 crc kubenswrapper[4725]: I1202 13:08:37.060625 4725 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:37 crc kubenswrapper[4725]: I1202 13:08:37.060866 4725 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:37 crc kubenswrapper[4725]: I1202 13:08:37.064042 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:38 crc kubenswrapper[4725]: I1202 13:08:38.064324 4725 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:38 crc kubenswrapper[4725]: I1202 13:08:38.064351 4725 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5b3cacb-78af-4fb7-a149-9d500e22828f" Dec 02 13:08:39 crc kubenswrapper[4725]: I1202 13:08:39.284776 4725 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f0a7843d-806b-47d6-99cc-35fb18f99b23" Dec 02 13:08:40 crc kubenswrapper[4725]: I1202 13:08:40.597256 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:08:41 crc kubenswrapper[4725]: I1202 13:08:41.888641 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:08:41 crc kubenswrapper[4725]: I1202 13:08:41.888885 4725 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 02 13:08:41 crc kubenswrapper[4725]: I1202 13:08:41.889199 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 02 13:08:43 crc kubenswrapper[4725]: I1202 13:08:43.119811 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 02 13:08:43 crc kubenswrapper[4725]: I1202 13:08:43.843953 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 13:08:45 crc kubenswrapper[4725]: I1202 13:08:45.593354 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 02 13:08:48 crc kubenswrapper[4725]: I1202 13:08:48.209684 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 02 13:08:48 crc kubenswrapper[4725]: I1202 13:08:48.328022 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 02 13:08:48 crc kubenswrapper[4725]: I1202 13:08:48.939951 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.202918 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.281189 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.397581 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.571666 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.615620 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.701883 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.790652 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 02 13:08:49 crc kubenswrapper[4725]: I1202 13:08:49.850005 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.041916 4725 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.179739 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.483178 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.530962 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.534305 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.608203 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.678738 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.733351 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.762965 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.816772 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.887951 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 02 13:08:50 crc kubenswrapper[4725]: I1202 13:08:50.903848 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.038951 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.103254 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.108736 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.297015 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.528509 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.539384 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.676786 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.680103 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.745696 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.748186 4725 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.827794 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.887952 4725 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.888017 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.941045 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.941681 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 02 13:08:51 crc kubenswrapper[4725]: I1202 13:08:51.995441 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.061692 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.089698 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.140952 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.216230 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.289646 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.310136 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.351772 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.394866 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.480829 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.542593 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.556848 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.628098 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.683506 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.703990 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.720839 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.902195 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.954443 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 02 13:08:52 crc kubenswrapper[4725]: I1202 13:08:52.971853 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.122173 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.133356 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.233306 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.254121 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.258069 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.354582 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.357879 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.383621 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.393673 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.404065 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.497098 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 02 13:08:53 crc kubenswrapper[4725]: I1202 13:08:53.566802 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.086487 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.086784 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.086870 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.093845 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.093983 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.094136 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.094189 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.094433 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.094560 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.098068 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.127966 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.133700 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.164033 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.280358 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.316975 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.459070 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.519986 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.565364 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.602756 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.614548 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.615353 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.639984 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.712836 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.799034 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.824191 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 02 13:08:54 crc kubenswrapper[4725]: I1202 13:08:54.877783 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.217044 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.265875 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.339808 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.379591 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.492119 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.593883 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.609179 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.615301 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.618739 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.659512 4725 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.761390 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.802264 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 02 13:08:55 crc kubenswrapper[4725]: I1202 13:08:55.967231 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.015338 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.054230 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.085740 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.136570 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.136573 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.196489 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.226055 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.272776 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.291399 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.302859 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.306431 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.349709 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.422612 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.425321 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.491189 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.513615 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.535952 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.573978 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.646242 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.699523 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.760513 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.790588 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.802262 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.830287 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.839250 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.895694 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.947075 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.957799 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 02 13:08:56 crc kubenswrapper[4725]: I1202 13:08:56.960415 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.003317 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.129430 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.131232 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.214817 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.229935 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.241302 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.327829 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.403852 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.437219 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.454942 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.507175 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.545677 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.567313 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.585659 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.586902 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.606455 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.713164 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.787606 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.874701 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.885586 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 02 13:08:57 crc kubenswrapper[4725]: I1202 13:08:57.904869 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.009809 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.103194 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.126046 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.182854 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.246311 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.268613 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.269333 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.296590 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.329057 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.447442 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.467897 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.523822 4725 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.527076 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.527061572 podStartE2EDuration="41.527061572s" podCreationTimestamp="2025-12-02 13:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:08:37.016755773 +0000 UTC m=+247.973397468" watchObservedRunningTime="2025-12-02 13:08:58.527061572 +0000 UTC m=+269.483703257" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.528281 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.528328 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.532603 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 13:08:58 crc kubenswrapper[4725]: I1202 13:08:58.545355 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.545338722 podStartE2EDuration="22.545338722s" podCreationTimestamp="2025-12-02 13:08:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:08:58.543246079 +0000 UTC m=+269.499887784" watchObservedRunningTime="2025-12-02 13:08:58.545338722 +0000 UTC m=+269.501980427" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.012220 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.018477 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.099582 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.116885 4725 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.190157 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.208051 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.246835 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.273615 4725 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.273845 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e" gracePeriod=5 Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.294324 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.310327 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.495623 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.559029 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.563244 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.706142 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.725720 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.814058 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.840559 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.862573 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 02 13:08:59 crc kubenswrapper[4725]: I1202 13:08:59.963651 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.039832 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.106800 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.128266 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.157666 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.248314 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.291742 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.296889 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.298740 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.436138 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.476216 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.645645 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.721810 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.748583 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.763359 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.777348 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.793942 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.855544 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.870451 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.960813 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.963958 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 02 13:09:00 crc kubenswrapper[4725]: I1202 13:09:00.982374 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.000098 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.066140 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.244210 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.315619 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.363375 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.367989 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.435806 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.551042 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.601820 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.632815 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.642411 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.705501 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.714546 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.736105 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.752225 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.797534 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.877485 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.888452 4725 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.888696 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.888838 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.889390 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"f91b606eb7ca216cf92a16ebc9fd6089d8ffd36dc3c8f5e2daec88673cfec644"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.889589 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://f91b606eb7ca216cf92a16ebc9fd6089d8ffd36dc3c8f5e2daec88673cfec644" gracePeriod=30 Dec 02 13:09:01 crc kubenswrapper[4725]: I1202 13:09:01.967855 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.100130 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.212213 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.243489 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.342773 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.391959 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.475940 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.679709 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.706418 4725 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.714588 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 13:09:02 crc kubenswrapper[4725]: I1202 13:09:02.806696 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 02 13:09:03 crc kubenswrapper[4725]: I1202 13:09:03.103410 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 02 13:09:03 crc kubenswrapper[4725]: I1202 13:09:03.237033 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 02 13:09:03 crc kubenswrapper[4725]: I1202 13:09:03.307271 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 02 13:09:03 crc kubenswrapper[4725]: I1202 13:09:03.480156 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 02 13:09:03 crc kubenswrapper[4725]: I1202 13:09:03.607831 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 02 13:09:03 crc kubenswrapper[4725]: I1202 13:09:03.675068 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.223279 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.271890 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.360570 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.363917 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.716986 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.840521 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.840597 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972479 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972549 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972596 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972628 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972669 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972744 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972813 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972770 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.972855 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.973022 4725 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.973038 4725 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.973051 4725 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.973062 4725 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:04 crc kubenswrapper[4725]: I1202 13:09:04.982022 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.074142 4725 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.201893 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.201934 4725 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e" exitCode=137 Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.201972 4725 scope.go:117] "RemoveContainer" containerID="a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.202082 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.218059 4725 scope.go:117] "RemoveContainer" containerID="a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e" Dec 02 13:09:05 crc kubenswrapper[4725]: E1202 13:09:05.218422 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e\": container with ID starting with a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e not found: ID does not exist" containerID="a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.218473 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e"} err="failed to get container status \"a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e\": rpc error: code = NotFound desc = could not find container \"a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e\": container with ID starting with a93b390d15fdee38a132eb4987cbc2801ef2fdf0ecfad4c85d5fd0fc2240330e not found: ID does not exist" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.273868 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.274108 4725 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.283834 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.283880 4725 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="25cd6eab-4673-483d-adb1-6c62354e5560" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.286705 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.286733 4725 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="25cd6eab-4673-483d-adb1-6c62354e5560" Dec 02 13:09:05 crc kubenswrapper[4725]: I1202 13:09:05.512039 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 02 13:09:06 crc kubenswrapper[4725]: I1202 13:09:06.975575 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.315284 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w4rh7"] Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.315546 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w4rh7" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="registry-server" containerID="cri-o://11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa" gracePeriod=30 Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.324534 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntm7v"] Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.324802 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ntm7v" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="registry-server" containerID="cri-o://78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971" gracePeriod=30 Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.331810 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldhmm"] Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.332079 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerName="marketplace-operator" containerID="cri-o://edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299" gracePeriod=30 Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.336791 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5gxkr"] Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.337051 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5gxkr" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="registry-server" containerID="cri-o://8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995" gracePeriod=30 Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.342616 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42pds"] Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.342994 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-42pds" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="registry-server" containerID="cri-o://82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977" gracePeriod=30 Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.352152 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.829220 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.834152 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.848061 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.853050 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.857582 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.911042 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-operator-metrics\") pod \"44387809-9d4a-43c6-b61d-364ab1b6c48f\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.911081 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-utilities\") pod \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.911101 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-catalog-content\") pod \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.911145 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-utilities\") pod \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.911171 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-catalog-content\") pod \"362e7a06-0d49-46c4-b7e9-1bfc52125374\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.911192 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-utilities\") pod \"362e7a06-0d49-46c4-b7e9-1bfc52125374\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.912236 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-utilities" (OuterVolumeSpecName: "utilities") pod "362e7a06-0d49-46c4-b7e9-1bfc52125374" (UID: "362e7a06-0d49-46c4-b7e9-1bfc52125374"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.912257 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-utilities" (OuterVolumeSpecName: "utilities") pod "1da4e04b-0c48-48ed-b6cb-b1f2e563e800" (UID: "1da4e04b-0c48-48ed-b6cb-b1f2e563e800"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.917192 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "44387809-9d4a-43c6-b61d-364ab1b6c48f" (UID: "44387809-9d4a-43c6-b61d-364ab1b6c48f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.932713 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-utilities" (OuterVolumeSpecName: "utilities") pod "9517f12e-2c7b-4c89-b65d-a9412e82e9a1" (UID: "9517f12e-2c7b-4c89-b65d-a9412e82e9a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.962777 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1da4e04b-0c48-48ed-b6cb-b1f2e563e800" (UID: "1da4e04b-0c48-48ed-b6cb-b1f2e563e800"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:07 crc kubenswrapper[4725]: I1202 13:09:07.963307 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "362e7a06-0d49-46c4-b7e9-1bfc52125374" (UID: "362e7a06-0d49-46c4-b7e9-1bfc52125374"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.011805 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xkqg\" (UniqueName: \"kubernetes.io/projected/c3c613fb-ab24-4568-a644-33d9e06063c9-kube-api-access-5xkqg\") pod \"c3c613fb-ab24-4568-a644-33d9e06063c9\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.011864 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-catalog-content\") pod \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.011891 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl5lr\" (UniqueName: \"kubernetes.io/projected/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-kube-api-access-fl5lr\") pod \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\" (UID: \"1da4e04b-0c48-48ed-b6cb-b1f2e563e800\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.011925 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-catalog-content\") pod \"c3c613fb-ab24-4568-a644-33d9e06063c9\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.011950 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-utilities\") pod \"c3c613fb-ab24-4568-a644-33d9e06063c9\" (UID: \"c3c613fb-ab24-4568-a644-33d9e06063c9\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.011980 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmvm6\" (UniqueName: \"kubernetes.io/projected/362e7a06-0d49-46c4-b7e9-1bfc52125374-kube-api-access-zmvm6\") pod \"362e7a06-0d49-46c4-b7e9-1bfc52125374\" (UID: \"362e7a06-0d49-46c4-b7e9-1bfc52125374\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.012040 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-trusted-ca\") pod \"44387809-9d4a-43c6-b61d-364ab1b6c48f\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.012077 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqh9b\" (UniqueName: \"kubernetes.io/projected/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-kube-api-access-nqh9b\") pod \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\" (UID: \"9517f12e-2c7b-4c89-b65d-a9412e82e9a1\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.012107 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6d2x\" (UniqueName: \"kubernetes.io/projected/44387809-9d4a-43c6-b61d-364ab1b6c48f-kube-api-access-x6d2x\") pod \"44387809-9d4a-43c6-b61d-364ab1b6c48f\" (UID: \"44387809-9d4a-43c6-b61d-364ab1b6c48f\") " Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.012860 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-utilities" (OuterVolumeSpecName: "utilities") pod "c3c613fb-ab24-4568-a644-33d9e06063c9" (UID: "c3c613fb-ab24-4568-a644-33d9e06063c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.012923 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "44387809-9d4a-43c6-b61d-364ab1b6c48f" (UID: "44387809-9d4a-43c6-b61d-364ab1b6c48f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013144 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013158 4725 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013171 4725 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/44387809-9d4a-43c6-b61d-364ab1b6c48f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013182 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013192 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013203 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013213 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.013223 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/362e7a06-0d49-46c4-b7e9-1bfc52125374-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.015140 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362e7a06-0d49-46c4-b7e9-1bfc52125374-kube-api-access-zmvm6" (OuterVolumeSpecName: "kube-api-access-zmvm6") pod "362e7a06-0d49-46c4-b7e9-1bfc52125374" (UID: "362e7a06-0d49-46c4-b7e9-1bfc52125374"). InnerVolumeSpecName "kube-api-access-zmvm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.015261 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-kube-api-access-nqh9b" (OuterVolumeSpecName: "kube-api-access-nqh9b") pod "9517f12e-2c7b-4c89-b65d-a9412e82e9a1" (UID: "9517f12e-2c7b-4c89-b65d-a9412e82e9a1"). InnerVolumeSpecName "kube-api-access-nqh9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.015501 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-kube-api-access-fl5lr" (OuterVolumeSpecName: "kube-api-access-fl5lr") pod "1da4e04b-0c48-48ed-b6cb-b1f2e563e800" (UID: "1da4e04b-0c48-48ed-b6cb-b1f2e563e800"). InnerVolumeSpecName "kube-api-access-fl5lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.016060 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44387809-9d4a-43c6-b61d-364ab1b6c48f-kube-api-access-x6d2x" (OuterVolumeSpecName: "kube-api-access-x6d2x") pod "44387809-9d4a-43c6-b61d-364ab1b6c48f" (UID: "44387809-9d4a-43c6-b61d-364ab1b6c48f"). InnerVolumeSpecName "kube-api-access-x6d2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.017193 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c613fb-ab24-4568-a644-33d9e06063c9-kube-api-access-5xkqg" (OuterVolumeSpecName: "kube-api-access-5xkqg") pod "c3c613fb-ab24-4568-a644-33d9e06063c9" (UID: "c3c613fb-ab24-4568-a644-33d9e06063c9"). InnerVolumeSpecName "kube-api-access-5xkqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.031143 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9517f12e-2c7b-4c89-b65d-a9412e82e9a1" (UID: "9517f12e-2c7b-4c89-b65d-a9412e82e9a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.113959 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmvm6\" (UniqueName: \"kubernetes.io/projected/362e7a06-0d49-46c4-b7e9-1bfc52125374-kube-api-access-zmvm6\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.113992 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqh9b\" (UniqueName: \"kubernetes.io/projected/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-kube-api-access-nqh9b\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.114002 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6d2x\" (UniqueName: \"kubernetes.io/projected/44387809-9d4a-43c6-b61d-364ab1b6c48f-kube-api-access-x6d2x\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.114013 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xkqg\" (UniqueName: \"kubernetes.io/projected/c3c613fb-ab24-4568-a644-33d9e06063c9-kube-api-access-5xkqg\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.114021 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9517f12e-2c7b-4c89-b65d-a9412e82e9a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.114030 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl5lr\" (UniqueName: \"kubernetes.io/projected/1da4e04b-0c48-48ed-b6cb-b1f2e563e800-kube-api-access-fl5lr\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.116895 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3c613fb-ab24-4568-a644-33d9e06063c9" (UID: "c3c613fb-ab24-4568-a644-33d9e06063c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.214708 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c613fb-ab24-4568-a644-33d9e06063c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.219588 4725 generic.go:334] "Generic (PLEG): container finished" podID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerID="11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa" exitCode=0 Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.219629 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w4rh7" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.219657 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerDied","Data":"11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.219708 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w4rh7" event={"ID":"1da4e04b-0c48-48ed-b6cb-b1f2e563e800","Type":"ContainerDied","Data":"664cd16e771996c495d4ec7ba5d4f78b46f545c51c2cd6bc651a6c5a76353925"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.219730 4725 scope.go:117] "RemoveContainer" containerID="11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.221920 4725 generic.go:334] "Generic (PLEG): container finished" podID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerID="edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299" exitCode=0 Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.222011 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.222076 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" event={"ID":"44387809-9d4a-43c6-b61d-364ab1b6c48f","Type":"ContainerDied","Data":"edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.222104 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ldhmm" event={"ID":"44387809-9d4a-43c6-b61d-364ab1b6c48f","Type":"ContainerDied","Data":"157d980855f44a64e8f9d0f0e12bebacef5380aff82d60d466efa349363390e5"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.224441 4725 generic.go:334] "Generic (PLEG): container finished" podID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerID="82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977" exitCode=0 Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.224507 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42pds" event={"ID":"c3c613fb-ab24-4568-a644-33d9e06063c9","Type":"ContainerDied","Data":"82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.224526 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42pds" event={"ID":"c3c613fb-ab24-4568-a644-33d9e06063c9","Type":"ContainerDied","Data":"b37fe8bed41cb0df89d5e42e0d3260c97f0d676c810d97ab70e56edd9957dd69"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.224592 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42pds" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.233396 4725 generic.go:334] "Generic (PLEG): container finished" podID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerID="78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971" exitCode=0 Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.233433 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntm7v" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.233541 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntm7v" event={"ID":"362e7a06-0d49-46c4-b7e9-1bfc52125374","Type":"ContainerDied","Data":"78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.233574 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntm7v" event={"ID":"362e7a06-0d49-46c4-b7e9-1bfc52125374","Type":"ContainerDied","Data":"ca9d7dfb9dd0213f43ac64b52f396caf8f1297388dd7b14ef7cd208d2c403c9e"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.236923 4725 generic.go:334] "Generic (PLEG): container finished" podID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerID="8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995" exitCode=0 Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.236958 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5gxkr" event={"ID":"9517f12e-2c7b-4c89-b65d-a9412e82e9a1","Type":"ContainerDied","Data":"8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.236983 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5gxkr" event={"ID":"9517f12e-2c7b-4c89-b65d-a9412e82e9a1","Type":"ContainerDied","Data":"f870921e127edbb57cf2ccd1023032e9988196dbd06fb773b65825cfde7d9045"} Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.236997 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5gxkr" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.238420 4725 scope.go:117] "RemoveContainer" containerID="45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.248675 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w4rh7"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.252520 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w4rh7"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.261174 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42pds"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.265085 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-42pds"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.276345 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldhmm"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.279571 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ldhmm"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.282273 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ntm7v"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.283557 4725 scope.go:117] "RemoveContainer" containerID="06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.289503 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ntm7v"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.297497 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5gxkr"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.304900 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5gxkr"] Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.305160 4725 scope.go:117] "RemoveContainer" containerID="11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.306791 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa\": container with ID starting with 11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa not found: ID does not exist" containerID="11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.306824 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa"} err="failed to get container status \"11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa\": rpc error: code = NotFound desc = could not find container \"11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa\": container with ID starting with 11094f8eed0b44f9837710390347203fa2ff34b4ccd190626436b12d8c7f9dfa not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.306845 4725 scope.go:117] "RemoveContainer" containerID="45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.307068 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d\": container with ID starting with 45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d not found: ID does not exist" containerID="45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.307086 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d"} err="failed to get container status \"45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d\": rpc error: code = NotFound desc = could not find container \"45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d\": container with ID starting with 45f82bfef67b0358d3c4d579f1e5f9d0c96e735fa1d0cfc7cb6d9f19a071424d not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.307099 4725 scope.go:117] "RemoveContainer" containerID="06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.307329 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027\": container with ID starting with 06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027 not found: ID does not exist" containerID="06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.307348 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027"} err="failed to get container status \"06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027\": rpc error: code = NotFound desc = could not find container \"06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027\": container with ID starting with 06516807e9934c7389d5e569c088097551dc87a22a83dc9c85217ebf0571a027 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.307360 4725 scope.go:117] "RemoveContainer" containerID="edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.320929 4725 scope.go:117] "RemoveContainer" containerID="edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.321228 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299\": container with ID starting with edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299 not found: ID does not exist" containerID="edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.321261 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299"} err="failed to get container status \"edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299\": rpc error: code = NotFound desc = could not find container \"edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299\": container with ID starting with edf238a52b690fd7bc994a4fb4ff8bcc104d47a0ec0a62e9d500e1f0026db299 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.321284 4725 scope.go:117] "RemoveContainer" containerID="82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.338974 4725 scope.go:117] "RemoveContainer" containerID="5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.354195 4725 scope.go:117] "RemoveContainer" containerID="6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.368085 4725 scope.go:117] "RemoveContainer" containerID="82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.368604 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977\": container with ID starting with 82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977 not found: ID does not exist" containerID="82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.368660 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977"} err="failed to get container status \"82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977\": rpc error: code = NotFound desc = could not find container \"82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977\": container with ID starting with 82a369217a474c436e5d247a4247265056891a74fd153aed63fee6a495e0b977 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.368691 4725 scope.go:117] "RemoveContainer" containerID="5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.368990 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a\": container with ID starting with 5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a not found: ID does not exist" containerID="5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.369021 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a"} err="failed to get container status \"5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a\": rpc error: code = NotFound desc = could not find container \"5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a\": container with ID starting with 5548890f252347b56b84a811f81fc95a989bc5a9d659472365bf557a94df9d2a not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.369043 4725 scope.go:117] "RemoveContainer" containerID="6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.369318 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a\": container with ID starting with 6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a not found: ID does not exist" containerID="6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.369442 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a"} err="failed to get container status \"6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a\": rpc error: code = NotFound desc = could not find container \"6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a\": container with ID starting with 6441fee80751240e36f58fd736e7fef558b61e4871f3b5e9e52f1decf4f2b85a not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.369539 4725 scope.go:117] "RemoveContainer" containerID="78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.380732 4725 scope.go:117] "RemoveContainer" containerID="cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.392779 4725 scope.go:117] "RemoveContainer" containerID="0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.403989 4725 scope.go:117] "RemoveContainer" containerID="78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.404380 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971\": container with ID starting with 78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971 not found: ID does not exist" containerID="78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.404508 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971"} err="failed to get container status \"78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971\": rpc error: code = NotFound desc = could not find container \"78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971\": container with ID starting with 78a299e5c6ef5061a109f93fb218e8220ba1d101ffe075a0be397dc7f6d74971 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.404608 4725 scope.go:117] "RemoveContainer" containerID="cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.404896 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0\": container with ID starting with cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0 not found: ID does not exist" containerID="cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.405005 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0"} err="failed to get container status \"cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0\": rpc error: code = NotFound desc = could not find container \"cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0\": container with ID starting with cbc5859a47c7532fdda23fea6bd8a7d76e43045744387aca4c1bc6305d711ff0 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.405078 4725 scope.go:117] "RemoveContainer" containerID="0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.405502 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91\": container with ID starting with 0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91 not found: ID does not exist" containerID="0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.405612 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91"} err="failed to get container status \"0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91\": rpc error: code = NotFound desc = could not find container \"0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91\": container with ID starting with 0684a1d08e4085fcdc45e3bbbeba87272c38346327feffcfb85e73c0e02bef91 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.405720 4725 scope.go:117] "RemoveContainer" containerID="8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.417071 4725 scope.go:117] "RemoveContainer" containerID="ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.428061 4725 scope.go:117] "RemoveContainer" containerID="0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.476068 4725 scope.go:117] "RemoveContainer" containerID="8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.476731 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995\": container with ID starting with 8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995 not found: ID does not exist" containerID="8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.476887 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995"} err="failed to get container status \"8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995\": rpc error: code = NotFound desc = could not find container \"8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995\": container with ID starting with 8e1ef69501d64e041ae7168049136a056a42bde36a181fd5c62636ef89cd0995 not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.476992 4725 scope.go:117] "RemoveContainer" containerID="ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.477447 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c\": container with ID starting with ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c not found: ID does not exist" containerID="ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.477496 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c"} err="failed to get container status \"ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c\": rpc error: code = NotFound desc = could not find container \"ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c\": container with ID starting with ec5a7656ce182d952a2373e731d288c954ba64caf65c9faae8a7dc9adcc5108c not found: ID does not exist" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.477520 4725 scope.go:117] "RemoveContainer" containerID="0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c" Dec 02 13:09:08 crc kubenswrapper[4725]: E1202 13:09:08.477896 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c\": container with ID starting with 0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c not found: ID does not exist" containerID="0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c" Dec 02 13:09:08 crc kubenswrapper[4725]: I1202 13:09:08.477923 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c"} err="failed to get container status \"0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c\": rpc error: code = NotFound desc = could not find container \"0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c\": container with ID starting with 0b655cfa460abc14d76b8678250ba312fc0f04a747c50fa43e0e00260e71dd6c not found: ID does not exist" Dec 02 13:09:09 crc kubenswrapper[4725]: I1202 13:09:09.274215 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" path="/var/lib/kubelet/pods/1da4e04b-0c48-48ed-b6cb-b1f2e563e800/volumes" Dec 02 13:09:09 crc kubenswrapper[4725]: I1202 13:09:09.274904 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" path="/var/lib/kubelet/pods/362e7a06-0d49-46c4-b7e9-1bfc52125374/volumes" Dec 02 13:09:09 crc kubenswrapper[4725]: I1202 13:09:09.275514 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" path="/var/lib/kubelet/pods/44387809-9d4a-43c6-b61d-364ab1b6c48f/volumes" Dec 02 13:09:09 crc kubenswrapper[4725]: I1202 13:09:09.275960 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" path="/var/lib/kubelet/pods/9517f12e-2c7b-4c89-b65d-a9412e82e9a1/volumes" Dec 02 13:09:09 crc kubenswrapper[4725]: I1202 13:09:09.276589 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" path="/var/lib/kubelet/pods/c3c613fb-ab24-4568-a644-33d9e06063c9/volumes" Dec 02 13:09:33 crc kubenswrapper[4725]: I1202 13:09:33.353140 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 02 13:09:33 crc kubenswrapper[4725]: I1202 13:09:33.355540 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 13:09:33 crc kubenswrapper[4725]: I1202 13:09:33.355807 4725 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f91b606eb7ca216cf92a16ebc9fd6089d8ffd36dc3c8f5e2daec88673cfec644" exitCode=137 Dec 02 13:09:33 crc kubenswrapper[4725]: I1202 13:09:33.355836 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f91b606eb7ca216cf92a16ebc9fd6089d8ffd36dc3c8f5e2daec88673cfec644"} Dec 02 13:09:33 crc kubenswrapper[4725]: I1202 13:09:33.355861 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"39b56cb5e8bc95074abfa0522a34370ccd9d2bd802339f453806b9940fa964bf"} Dec 02 13:09:33 crc kubenswrapper[4725]: I1202 13:09:33.355876 4725 scope.go:117] "RemoveContainer" containerID="ea3a093d237349cdd478143b1ee2327586b2d4d9199f62953de18db03aae790f" Dec 02 13:09:34 crc kubenswrapper[4725]: I1202 13:09:34.362600 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 02 13:09:40 crc kubenswrapper[4725]: I1202 13:09:40.597088 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:09:41 crc kubenswrapper[4725]: I1202 13:09:41.888473 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:09:41 crc kubenswrapper[4725]: I1202 13:09:41.891905 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:09:50 crc kubenswrapper[4725]: I1202 13:09:50.601198 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.617497 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt"] Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.617989 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" podUID="4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" containerName="route-controller-manager" containerID="cri-o://12fd158fb839f459bde6993a22ca260a336cbb67a03efc2cd8c905805c7fb705" gracePeriod=30 Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.630663 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kcshx"] Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.631200 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.631319 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.631417 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.631529 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.631619 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.631704 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.631789 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.631867 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.631955 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.632037 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.632120 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.632205 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.632292 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerName="marketplace-operator" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.632374 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerName="marketplace-operator" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.632451 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.632551 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.632629 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.633283 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.633375 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.633454 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.633564 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.633639 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="extract-utilities" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.633736 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.633816 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.633893 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.633980 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.634075 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="276b30df-586f-4eba-a848-5b4230667534" containerName="installer" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.634155 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="276b30df-586f-4eba-a848-5b4230667534" containerName="installer" Dec 02 13:09:51 crc kubenswrapper[4725]: E1202 13:09:51.634239 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.634325 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="extract-content" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.634787 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.634891 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="44387809-9d4a-43c6-b61d-364ab1b6c48f" containerName="marketplace-operator" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.634984 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da4e04b-0c48-48ed-b6cb-b1f2e563e800" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.635092 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="9517f12e-2c7b-4c89-b65d-a9412e82e9a1" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.635175 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="362e7a06-0d49-46c4-b7e9-1bfc52125374" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.635268 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="276b30df-586f-4eba-a848-5b4230667534" containerName="installer" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.635365 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c613fb-ab24-4568-a644-33d9e06063c9" containerName="registry-server" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.635872 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q9s7m"] Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.636049 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.636268 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" podUID="62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" containerName="controller-manager" containerID="cri-o://e9291ff65ad0c3223ee71dddffee9aa2e7373a035281fc3a68d9aacd8642b1f8" gracePeriod=30 Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.650149 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.650198 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.651519 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.651664 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.654318 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.655483 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kcshx"] Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.709779 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55b1c898-ab77-4a63-ba54-4c3246ec5732-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.709903 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-798cz\" (UniqueName: \"kubernetes.io/projected/55b1c898-ab77-4a63-ba54-4c3246ec5732-kube-api-access-798cz\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.709998 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55b1c898-ab77-4a63-ba54-4c3246ec5732-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.810848 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-798cz\" (UniqueName: \"kubernetes.io/projected/55b1c898-ab77-4a63-ba54-4c3246ec5732-kube-api-access-798cz\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.810892 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55b1c898-ab77-4a63-ba54-4c3246ec5732-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.810934 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55b1c898-ab77-4a63-ba54-4c3246ec5732-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.812148 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55b1c898-ab77-4a63-ba54-4c3246ec5732-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.818815 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55b1c898-ab77-4a63-ba54-4c3246ec5732-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.832197 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-798cz\" (UniqueName: \"kubernetes.io/projected/55b1c898-ab77-4a63-ba54-4c3246ec5732-kube-api-access-798cz\") pod \"marketplace-operator-79b997595-kcshx\" (UID: \"55b1c898-ab77-4a63-ba54-4c3246ec5732\") " pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:51 crc kubenswrapper[4725]: I1202 13:09:51.965567 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:52 crc kubenswrapper[4725]: I1202 13:09:52.431951 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kcshx"] Dec 02 13:09:52 crc kubenswrapper[4725]: I1202 13:09:52.454641 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" event={"ID":"55b1c898-ab77-4a63-ba54-4c3246ec5732","Type":"ContainerStarted","Data":"6117e7cfbcf1e3a583bed06d0d51b0eeacf3ffc53dbdcf3c3f41f06c39887731"} Dec 02 13:09:52 crc kubenswrapper[4725]: I1202 13:09:52.456681 4725 generic.go:334] "Generic (PLEG): container finished" podID="4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" containerID="12fd158fb839f459bde6993a22ca260a336cbb67a03efc2cd8c905805c7fb705" exitCode=0 Dec 02 13:09:52 crc kubenswrapper[4725]: I1202 13:09:52.456847 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" event={"ID":"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f","Type":"ContainerDied","Data":"12fd158fb839f459bde6993a22ca260a336cbb67a03efc2cd8c905805c7fb705"} Dec 02 13:09:52 crc kubenswrapper[4725]: I1202 13:09:52.459199 4725 generic.go:334] "Generic (PLEG): container finished" podID="62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" containerID="e9291ff65ad0c3223ee71dddffee9aa2e7373a035281fc3a68d9aacd8642b1f8" exitCode=0 Dec 02 13:09:52 crc kubenswrapper[4725]: I1202 13:09:52.459247 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" event={"ID":"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911","Type":"ContainerDied","Data":"e9291ff65ad0c3223ee71dddffee9aa2e7373a035281fc3a68d9aacd8642b1f8"} Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.188951 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.221585 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h"] Dec 02 13:09:53 crc kubenswrapper[4725]: E1202 13:09:53.221819 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" containerName="route-controller-manager" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.221837 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" containerName="route-controller-manager" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.221934 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" containerName="route-controller-manager" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.222320 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.231394 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h"] Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.246341 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333031 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-config\") pod \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333120 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghk2k\" (UniqueName: \"kubernetes.io/projected/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-kube-api-access-ghk2k\") pod \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333167 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-client-ca\") pod \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333216 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-serving-cert\") pod \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\" (UID: \"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333543 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8523c6ac-24e9-4fad-acb5-852c252f40dc-serving-cert\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333582 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8523c6ac-24e9-4fad-acb5-852c252f40dc-client-ca\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333636 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8523c6ac-24e9-4fad-acb5-852c252f40dc-config\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.333752 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mgs7\" (UniqueName: \"kubernetes.io/projected/8523c6ac-24e9-4fad-acb5-852c252f40dc-kube-api-access-7mgs7\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.334040 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-client-ca" (OuterVolumeSpecName: "client-ca") pod "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" (UID: "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.334138 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-config" (OuterVolumeSpecName: "config") pod "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" (UID: "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.339401 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" (UID: "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.339978 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-kube-api-access-ghk2k" (OuterVolumeSpecName: "kube-api-access-ghk2k") pod "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" (UID: "4bbbf1bf-0627-4bbe-b158-2a889ab37f2f"). InnerVolumeSpecName "kube-api-access-ghk2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.434709 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-serving-cert\") pod \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.435902 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-config\") pod \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.435976 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-client-ca\") pod \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436051 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-proxy-ca-bundles\") pod \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436154 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46x5c\" (UniqueName: \"kubernetes.io/projected/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-kube-api-access-46x5c\") pod \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\" (UID: \"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911\") " Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436419 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8523c6ac-24e9-4fad-acb5-852c252f40dc-serving-cert\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436470 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8523c6ac-24e9-4fad-acb5-852c252f40dc-client-ca\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436519 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8523c6ac-24e9-4fad-acb5-852c252f40dc-config\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436555 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mgs7\" (UniqueName: \"kubernetes.io/projected/8523c6ac-24e9-4fad-acb5-852c252f40dc-kube-api-access-7mgs7\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436643 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436657 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghk2k\" (UniqueName: \"kubernetes.io/projected/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-kube-api-access-ghk2k\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436668 4725 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436680 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436756 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-client-ca" (OuterVolumeSpecName: "client-ca") pod "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" (UID: "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436770 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" (UID: "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.436824 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-config" (OuterVolumeSpecName: "config") pod "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" (UID: "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.437827 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8523c6ac-24e9-4fad-acb5-852c252f40dc-client-ca\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.438204 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" (UID: "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.438604 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8523c6ac-24e9-4fad-acb5-852c252f40dc-config\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.439287 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-kube-api-access-46x5c" (OuterVolumeSpecName: "kube-api-access-46x5c") pod "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" (UID: "62766e8b-0dd9-42e4-a1f1-4c74ae7f3911"). InnerVolumeSpecName "kube-api-access-46x5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.441572 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8523c6ac-24e9-4fad-acb5-852c252f40dc-serving-cert\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.453197 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mgs7\" (UniqueName: \"kubernetes.io/projected/8523c6ac-24e9-4fad-acb5-852c252f40dc-kube-api-access-7mgs7\") pod \"route-controller-manager-769b794f5d-6td5h\" (UID: \"8523c6ac-24e9-4fad-acb5-852c252f40dc\") " pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.466003 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" event={"ID":"55b1c898-ab77-4a63-ba54-4c3246ec5732","Type":"ContainerStarted","Data":"e049e9a4fb471c4121805b21694b2f221a7a72e3633e76715c6ce4ebb61c62f7"} Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.466161 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.468824 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" event={"ID":"4bbbf1bf-0627-4bbe-b158-2a889ab37f2f","Type":"ContainerDied","Data":"1b7ee3727b445416530c7dd0e6df25c401310e00b705e0708e2742e260ee0ff8"} Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.468885 4725 scope.go:117] "RemoveContainer" containerID="12fd158fb839f459bde6993a22ca260a336cbb67a03efc2cd8c905805c7fb705" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.469026 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.469318 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.472657 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" event={"ID":"62766e8b-0dd9-42e4-a1f1-4c74ae7f3911","Type":"ContainerDied","Data":"ad2a9c43de628a7e3017e5fa810ae1d4f7692f5cb145aef59d508b39a40ff84a"} Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.472766 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q9s7m" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.483116 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kcshx" podStartSLOduration=2.483097528 podStartE2EDuration="2.483097528s" podCreationTimestamp="2025-12-02 13:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:09:53.481804435 +0000 UTC m=+324.438446150" watchObservedRunningTime="2025-12-02 13:09:53.483097528 +0000 UTC m=+324.439739223" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.491041 4725 scope.go:117] "RemoveContainer" containerID="e9291ff65ad0c3223ee71dddffee9aa2e7373a035281fc3a68d9aacd8642b1f8" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.532652 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q9s7m"] Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.538327 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.538366 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.538378 4725 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.538395 4725 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.538412 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46x5c\" (UniqueName: \"kubernetes.io/projected/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911-kube-api-access-46x5c\") on node \"crc\" DevicePath \"\"" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.540038 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q9s7m"] Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.545890 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt"] Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.551835 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fqcrt"] Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.552017 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:53 crc kubenswrapper[4725]: I1202 13:09:53.944613 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h"] Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.284874 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.285764 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.478902 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" event={"ID":"8523c6ac-24e9-4fad-acb5-852c252f40dc","Type":"ContainerStarted","Data":"e3ef7ec9e5d0775ba265c1fb205840bec377ecc01139e22007098f9e554d7cbd"} Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.478935 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" event={"ID":"8523c6ac-24e9-4fad-acb5-852c252f40dc","Type":"ContainerStarted","Data":"2c9a1b968b1e1017590186384aba2ef3c5a127d252e96bac4b750a66a52c1f61"} Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.479040 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.484258 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" Dec 02 13:09:54 crc kubenswrapper[4725]: I1202 13:09:54.511788 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-769b794f5d-6td5h" podStartSLOduration=3.511770613 podStartE2EDuration="3.511770613s" podCreationTimestamp="2025-12-02 13:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:09:54.497112234 +0000 UTC m=+325.453753929" watchObservedRunningTime="2025-12-02 13:09:54.511770613 +0000 UTC m=+325.468412298" Dec 02 13:09:55 crc kubenswrapper[4725]: I1202 13:09:55.273576 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bbbf1bf-0627-4bbe-b158-2a889ab37f2f" path="/var/lib/kubelet/pods/4bbbf1bf-0627-4bbe-b158-2a889ab37f2f/volumes" Dec 02 13:09:55 crc kubenswrapper[4725]: I1202 13:09:55.274424 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" path="/var/lib/kubelet/pods/62766e8b-0dd9-42e4-a1f1-4c74ae7f3911/volumes" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.074976 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f6fc7c544-fcs86"] Dec 02 13:09:56 crc kubenswrapper[4725]: E1202 13:09:56.075164 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" containerName="controller-manager" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.075175 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" containerName="controller-manager" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.075262 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="62766e8b-0dd9-42e4-a1f1-4c74ae7f3911" containerName="controller-manager" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.075780 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.078755 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.079121 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.079141 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.079229 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.079344 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.081057 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.087718 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.093593 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f6fc7c544-fcs86"] Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.183746 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7p49\" (UniqueName: \"kubernetes.io/projected/13a78439-0d99-4068-8a03-cd888dd1249b-kube-api-access-v7p49\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.183810 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-client-ca\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.183887 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13a78439-0d99-4068-8a03-cd888dd1249b-serving-cert\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.183911 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-config\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.183940 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-proxy-ca-bundles\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.285397 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13a78439-0d99-4068-8a03-cd888dd1249b-serving-cert\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.285532 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-config\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.285569 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-proxy-ca-bundles\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.285614 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7p49\" (UniqueName: \"kubernetes.io/projected/13a78439-0d99-4068-8a03-cd888dd1249b-kube-api-access-v7p49\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.285658 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-client-ca\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.286666 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-client-ca\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.286980 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-config\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.287008 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-proxy-ca-bundles\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.290698 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13a78439-0d99-4068-8a03-cd888dd1249b-serving-cert\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.311279 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7p49\" (UniqueName: \"kubernetes.io/projected/13a78439-0d99-4068-8a03-cd888dd1249b-kube-api-access-v7p49\") pod \"controller-manager-6f6fc7c544-fcs86\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.389383 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:56 crc kubenswrapper[4725]: I1202 13:09:56.541680 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f6fc7c544-fcs86"] Dec 02 13:09:56 crc kubenswrapper[4725]: W1202 13:09:56.547201 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13a78439_0d99_4068_8a03_cd888dd1249b.slice/crio-8f2be59e927d6e25b99b5559d2ac9224fd183037378cd1c6dd3ae750fa503487 WatchSource:0}: Error finding container 8f2be59e927d6e25b99b5559d2ac9224fd183037378cd1c6dd3ae750fa503487: Status 404 returned error can't find the container with id 8f2be59e927d6e25b99b5559d2ac9224fd183037378cd1c6dd3ae750fa503487 Dec 02 13:09:57 crc kubenswrapper[4725]: I1202 13:09:57.500999 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" event={"ID":"13a78439-0d99-4068-8a03-cd888dd1249b","Type":"ContainerStarted","Data":"170afded2c6443277e7d82a79804cea82fce5fd8f79fe3b6481a2e611fecffac"} Dec 02 13:09:57 crc kubenswrapper[4725]: I1202 13:09:57.501324 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" event={"ID":"13a78439-0d99-4068-8a03-cd888dd1249b","Type":"ContainerStarted","Data":"8f2be59e927d6e25b99b5559d2ac9224fd183037378cd1c6dd3ae750fa503487"} Dec 02 13:09:57 crc kubenswrapper[4725]: I1202 13:09:57.501339 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:57 crc kubenswrapper[4725]: I1202 13:09:57.505388 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:09:57 crc kubenswrapper[4725]: I1202 13:09:57.520375 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" podStartSLOduration=6.520355319 podStartE2EDuration="6.520355319s" podCreationTimestamp="2025-12-02 13:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:09:57.517097738 +0000 UTC m=+328.473739433" watchObservedRunningTime="2025-12-02 13:09:57.520355319 +0000 UTC m=+328.476997004" Dec 02 13:09:58 crc kubenswrapper[4725]: I1202 13:09:58.853317 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8mhmp"] Dec 02 13:09:58 crc kubenswrapper[4725]: I1202 13:09:58.854530 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:58 crc kubenswrapper[4725]: I1202 13:09:58.859700 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 13:09:58 crc kubenswrapper[4725]: I1202 13:09:58.859918 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8mhmp"] Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.015319 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/1ef8608f-fa22-48e6-858b-c23693d8133e-kube-api-access-x5xkd\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.015384 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-utilities\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.015421 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-catalog-content\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.042180 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ss4h6"] Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.043596 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.045954 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.051444 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ss4h6"] Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.116871 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-catalog-content\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.116962 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/1ef8608f-fa22-48e6-858b-c23693d8133e-kube-api-access-x5xkd\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.117010 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-utilities\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.117426 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-catalog-content\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.117476 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-utilities\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.134130 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/1ef8608f-fa22-48e6-858b-c23693d8133e-kube-api-access-x5xkd\") pod \"redhat-operators-8mhmp\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.171656 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.218449 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2ff71e-0534-496d-8c56-2413e0352da4-utilities\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.218824 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx8xf\" (UniqueName: \"kubernetes.io/projected/1c2ff71e-0534-496d-8c56-2413e0352da4-kube-api-access-dx8xf\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.218902 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2ff71e-0534-496d-8c56-2413e0352da4-catalog-content\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.320268 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2ff71e-0534-496d-8c56-2413e0352da4-catalog-content\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.320334 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2ff71e-0534-496d-8c56-2413e0352da4-utilities\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.320358 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx8xf\" (UniqueName: \"kubernetes.io/projected/1c2ff71e-0534-496d-8c56-2413e0352da4-kube-api-access-dx8xf\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.320962 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c2ff71e-0534-496d-8c56-2413e0352da4-catalog-content\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.321121 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c2ff71e-0534-496d-8c56-2413e0352da4-utilities\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.337206 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx8xf\" (UniqueName: \"kubernetes.io/projected/1c2ff71e-0534-496d-8c56-2413e0352da4-kube-api-access-dx8xf\") pod \"redhat-marketplace-ss4h6\" (UID: \"1c2ff71e-0534-496d-8c56-2413e0352da4\") " pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.357735 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.557164 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8mhmp"] Dec 02 13:09:59 crc kubenswrapper[4725]: W1202 13:09:59.560838 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ef8608f_fa22_48e6_858b_c23693d8133e.slice/crio-107f368fdf8f65459629fd0630dc265a6b348fcfdbb4b5c153e16aab09906c78 WatchSource:0}: Error finding container 107f368fdf8f65459629fd0630dc265a6b348fcfdbb4b5c153e16aab09906c78: Status 404 returned error can't find the container with id 107f368fdf8f65459629fd0630dc265a6b348fcfdbb4b5c153e16aab09906c78 Dec 02 13:09:59 crc kubenswrapper[4725]: I1202 13:09:59.749614 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ss4h6"] Dec 02 13:10:00 crc kubenswrapper[4725]: I1202 13:10:00.518356 4725 generic.go:334] "Generic (PLEG): container finished" podID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerID="60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b" exitCode=0 Dec 02 13:10:00 crc kubenswrapper[4725]: I1202 13:10:00.518434 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8mhmp" event={"ID":"1ef8608f-fa22-48e6-858b-c23693d8133e","Type":"ContainerDied","Data":"60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b"} Dec 02 13:10:00 crc kubenswrapper[4725]: I1202 13:10:00.518524 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8mhmp" event={"ID":"1ef8608f-fa22-48e6-858b-c23693d8133e","Type":"ContainerStarted","Data":"107f368fdf8f65459629fd0630dc265a6b348fcfdbb4b5c153e16aab09906c78"} Dec 02 13:10:00 crc kubenswrapper[4725]: I1202 13:10:00.520808 4725 generic.go:334] "Generic (PLEG): container finished" podID="1c2ff71e-0534-496d-8c56-2413e0352da4" containerID="6400ac9c8cdbe8427a1c0505b846e0cb89ea5a2bfeff7443f1b60342fef6ab33" exitCode=0 Dec 02 13:10:00 crc kubenswrapper[4725]: I1202 13:10:00.520837 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ss4h6" event={"ID":"1c2ff71e-0534-496d-8c56-2413e0352da4","Type":"ContainerDied","Data":"6400ac9c8cdbe8427a1c0505b846e0cb89ea5a2bfeff7443f1b60342fef6ab33"} Dec 02 13:10:00 crc kubenswrapper[4725]: I1202 13:10:00.520858 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ss4h6" event={"ID":"1c2ff71e-0534-496d-8c56-2413e0352da4","Type":"ContainerStarted","Data":"f50e0bb9320499cf988c84b2ad1c153ad4c48acc4c01c3f18ff4c957fb3f1ee9"} Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.443212 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rs8r5"] Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.444495 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.447284 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.454993 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rs8r5"] Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.526212 4725 generic.go:334] "Generic (PLEG): container finished" podID="1c2ff71e-0534-496d-8c56-2413e0352da4" containerID="36235a1fe15ce2d7b83b014f42077146fea453a30b9c8aef4ef7e85cee2d04e9" exitCode=0 Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.526257 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ss4h6" event={"ID":"1c2ff71e-0534-496d-8c56-2413e0352da4","Type":"ContainerDied","Data":"36235a1fe15ce2d7b83b014f42077146fea453a30b9c8aef4ef7e85cee2d04e9"} Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.550303 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d5c377-0464-41bd-aff4-de31874ea6b9-utilities\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.550399 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d5c377-0464-41bd-aff4-de31874ea6b9-catalog-content\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.550665 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d46bw\" (UniqueName: \"kubernetes.io/projected/38d5c377-0464-41bd-aff4-de31874ea6b9-kube-api-access-d46bw\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.647895 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sg828"] Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.651755 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.652047 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwkt\" (UniqueName: \"kubernetes.io/projected/d64eb153-6898-4b22-b86b-df77e2c71044-kube-api-access-mvwkt\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.652124 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d5c377-0464-41bd-aff4-de31874ea6b9-catalog-content\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.652173 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d46bw\" (UniqueName: \"kubernetes.io/projected/38d5c377-0464-41bd-aff4-de31874ea6b9-kube-api-access-d46bw\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.652203 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-catalog-content\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.652233 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d5c377-0464-41bd-aff4-de31874ea6b9-utilities\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.652249 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-utilities\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.653031 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d5c377-0464-41bd-aff4-de31874ea6b9-catalog-content\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.653053 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sg828"] Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.653270 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d5c377-0464-41bd-aff4-de31874ea6b9-utilities\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.653748 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.676927 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d46bw\" (UniqueName: \"kubernetes.io/projected/38d5c377-0464-41bd-aff4-de31874ea6b9-kube-api-access-d46bw\") pod \"certified-operators-rs8r5\" (UID: \"38d5c377-0464-41bd-aff4-de31874ea6b9\") " pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.753364 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwkt\" (UniqueName: \"kubernetes.io/projected/d64eb153-6898-4b22-b86b-df77e2c71044-kube-api-access-mvwkt\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.753583 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-catalog-content\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.753614 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-utilities\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.754087 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-utilities\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.754259 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-catalog-content\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.775379 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwkt\" (UniqueName: \"kubernetes.io/projected/d64eb153-6898-4b22-b86b-df77e2c71044-kube-api-access-mvwkt\") pod \"community-operators-sg828\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.821233 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:01 crc kubenswrapper[4725]: I1202 13:10:01.968027 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.214033 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rs8r5"] Dec 02 13:10:02 crc kubenswrapper[4725]: W1202 13:10:02.220765 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38d5c377_0464_41bd_aff4_de31874ea6b9.slice/crio-80a32a5604291a7856afa627dd8eb1ea74a14c5e64ad2503179409ffc61dca15 WatchSource:0}: Error finding container 80a32a5604291a7856afa627dd8eb1ea74a14c5e64ad2503179409ffc61dca15: Status 404 returned error can't find the container with id 80a32a5604291a7856afa627dd8eb1ea74a14c5e64ad2503179409ffc61dca15 Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.344764 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sg828"] Dec 02 13:10:02 crc kubenswrapper[4725]: W1202 13:10:02.425379 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd64eb153_6898_4b22_b86b_df77e2c71044.slice/crio-2aa0aaa86c9a44c555e96f8aa374e678a9d67e5045217d734386ff22f12e0228 WatchSource:0}: Error finding container 2aa0aaa86c9a44c555e96f8aa374e678a9d67e5045217d734386ff22f12e0228: Status 404 returned error can't find the container with id 2aa0aaa86c9a44c555e96f8aa374e678a9d67e5045217d734386ff22f12e0228 Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.533657 4725 generic.go:334] "Generic (PLEG): container finished" podID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerID="a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c" exitCode=0 Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.533751 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8mhmp" event={"ID":"1ef8608f-fa22-48e6-858b-c23693d8133e","Type":"ContainerDied","Data":"a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c"} Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.535054 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg828" event={"ID":"d64eb153-6898-4b22-b86b-df77e2c71044","Type":"ContainerStarted","Data":"2aa0aaa86c9a44c555e96f8aa374e678a9d67e5045217d734386ff22f12e0228"} Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.537024 4725 generic.go:334] "Generic (PLEG): container finished" podID="38d5c377-0464-41bd-aff4-de31874ea6b9" containerID="5570fb4714f807d6136e913d3e140a15bfa45e6fa0f418c1abb0772d35aad0b6" exitCode=0 Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.537064 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rs8r5" event={"ID":"38d5c377-0464-41bd-aff4-de31874ea6b9","Type":"ContainerDied","Data":"5570fb4714f807d6136e913d3e140a15bfa45e6fa0f418c1abb0772d35aad0b6"} Dec 02 13:10:02 crc kubenswrapper[4725]: I1202 13:10:02.537095 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rs8r5" event={"ID":"38d5c377-0464-41bd-aff4-de31874ea6b9","Type":"ContainerStarted","Data":"80a32a5604291a7856afa627dd8eb1ea74a14c5e64ad2503179409ffc61dca15"} Dec 02 13:10:03 crc kubenswrapper[4725]: I1202 13:10:03.545469 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8mhmp" event={"ID":"1ef8608f-fa22-48e6-858b-c23693d8133e","Type":"ContainerStarted","Data":"c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692"} Dec 02 13:10:03 crc kubenswrapper[4725]: I1202 13:10:03.546701 4725 generic.go:334] "Generic (PLEG): container finished" podID="d64eb153-6898-4b22-b86b-df77e2c71044" containerID="372daf2d90a65c03b4b0661fb65af4ba15d86e063c397b9921093d2e144df547" exitCode=0 Dec 02 13:10:03 crc kubenswrapper[4725]: I1202 13:10:03.546752 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg828" event={"ID":"d64eb153-6898-4b22-b86b-df77e2c71044","Type":"ContainerDied","Data":"372daf2d90a65c03b4b0661fb65af4ba15d86e063c397b9921093d2e144df547"} Dec 02 13:10:03 crc kubenswrapper[4725]: I1202 13:10:03.549670 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ss4h6" event={"ID":"1c2ff71e-0534-496d-8c56-2413e0352da4","Type":"ContainerStarted","Data":"eb3ca8398c9da3384a03351b7c129d349834558cce6835f27a0b581428a07a43"} Dec 02 13:10:03 crc kubenswrapper[4725]: I1202 13:10:03.564388 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8mhmp" podStartSLOduration=2.660024335 podStartE2EDuration="5.564368011s" podCreationTimestamp="2025-12-02 13:09:58 +0000 UTC" firstStartedPulling="2025-12-02 13:10:00.519798226 +0000 UTC m=+331.476439921" lastFinishedPulling="2025-12-02 13:10:03.424141892 +0000 UTC m=+334.380783597" observedRunningTime="2025-12-02 13:10:03.564017103 +0000 UTC m=+334.520658798" watchObservedRunningTime="2025-12-02 13:10:03.564368011 +0000 UTC m=+334.521009706" Dec 02 13:10:03 crc kubenswrapper[4725]: I1202 13:10:03.616329 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ss4h6" podStartSLOduration=2.453654064 podStartE2EDuration="4.61631461s" podCreationTimestamp="2025-12-02 13:09:59 +0000 UTC" firstStartedPulling="2025-12-02 13:10:00.522936873 +0000 UTC m=+331.479578568" lastFinishedPulling="2025-12-02 13:10:02.685597419 +0000 UTC m=+333.642239114" observedRunningTime="2025-12-02 13:10:03.615343286 +0000 UTC m=+334.571984981" watchObservedRunningTime="2025-12-02 13:10:03.61631461 +0000 UTC m=+334.572956305" Dec 02 13:10:04 crc kubenswrapper[4725]: I1202 13:10:04.558603 4725 generic.go:334] "Generic (PLEG): container finished" podID="38d5c377-0464-41bd-aff4-de31874ea6b9" containerID="8ac0a6a45864fb4d3e46f95c09c89fe1d66e76b2e2fd230e7f44e27d851d3f05" exitCode=0 Dec 02 13:10:04 crc kubenswrapper[4725]: I1202 13:10:04.558699 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rs8r5" event={"ID":"38d5c377-0464-41bd-aff4-de31874ea6b9","Type":"ContainerDied","Data":"8ac0a6a45864fb4d3e46f95c09c89fe1d66e76b2e2fd230e7f44e27d851d3f05"} Dec 02 13:10:05 crc kubenswrapper[4725]: I1202 13:10:05.566754 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rs8r5" event={"ID":"38d5c377-0464-41bd-aff4-de31874ea6b9","Type":"ContainerStarted","Data":"09987105e8e4eef11117e6f3b2bc1f3a3419b82af8eb5cbea2097d0cf8c1e898"} Dec 02 13:10:05 crc kubenswrapper[4725]: I1202 13:10:05.568406 4725 generic.go:334] "Generic (PLEG): container finished" podID="d64eb153-6898-4b22-b86b-df77e2c71044" containerID="f1041bd51ae6778c41539a821ad5e17054386be65b2136ddc793d2956a797a9f" exitCode=0 Dec 02 13:10:05 crc kubenswrapper[4725]: I1202 13:10:05.568440 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg828" event={"ID":"d64eb153-6898-4b22-b86b-df77e2c71044","Type":"ContainerDied","Data":"f1041bd51ae6778c41539a821ad5e17054386be65b2136ddc793d2956a797a9f"} Dec 02 13:10:05 crc kubenswrapper[4725]: I1202 13:10:05.585401 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rs8r5" podStartSLOduration=2.064395785 podStartE2EDuration="4.585380042s" podCreationTimestamp="2025-12-02 13:10:01 +0000 UTC" firstStartedPulling="2025-12-02 13:10:02.538337615 +0000 UTC m=+333.494979310" lastFinishedPulling="2025-12-02 13:10:05.059321872 +0000 UTC m=+336.015963567" observedRunningTime="2025-12-02 13:10:05.5852996 +0000 UTC m=+336.541941295" watchObservedRunningTime="2025-12-02 13:10:05.585380042 +0000 UTC m=+336.542021737" Dec 02 13:10:06 crc kubenswrapper[4725]: I1202 13:10:06.576409 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg828" event={"ID":"d64eb153-6898-4b22-b86b-df77e2c71044","Type":"ContainerStarted","Data":"45711932e9ff315ab326cc4df554f6f4aaac3b7181b337652230df4722f777f7"} Dec 02 13:10:06 crc kubenswrapper[4725]: I1202 13:10:06.595733 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sg828" podStartSLOduration=3.102604905 podStartE2EDuration="5.595714009s" podCreationTimestamp="2025-12-02 13:10:01 +0000 UTC" firstStartedPulling="2025-12-02 13:10:03.553946993 +0000 UTC m=+334.510588678" lastFinishedPulling="2025-12-02 13:10:06.047056087 +0000 UTC m=+337.003697782" observedRunningTime="2025-12-02 13:10:06.592997572 +0000 UTC m=+337.549639277" watchObservedRunningTime="2025-12-02 13:10:06.595714009 +0000 UTC m=+337.552355704" Dec 02 13:10:08 crc kubenswrapper[4725]: I1202 13:10:08.264091 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f6fc7c544-fcs86"] Dec 02 13:10:08 crc kubenswrapper[4725]: I1202 13:10:08.264291 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" podUID="13a78439-0d99-4068-8a03-cd888dd1249b" containerName="controller-manager" containerID="cri-o://170afded2c6443277e7d82a79804cea82fce5fd8f79fe3b6481a2e611fecffac" gracePeriod=30 Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.172906 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.173041 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.226998 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.358711 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.358749 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.394808 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.590687 4725 generic.go:334] "Generic (PLEG): container finished" podID="13a78439-0d99-4068-8a03-cd888dd1249b" containerID="170afded2c6443277e7d82a79804cea82fce5fd8f79fe3b6481a2e611fecffac" exitCode=0 Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.590764 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" event={"ID":"13a78439-0d99-4068-8a03-cd888dd1249b","Type":"ContainerDied","Data":"170afded2c6443277e7d82a79804cea82fce5fd8f79fe3b6481a2e611fecffac"} Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.624493 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.635304 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ss4h6" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.892539 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.918259 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l"] Dec 02 13:10:09 crc kubenswrapper[4725]: E1202 13:10:09.918520 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13a78439-0d99-4068-8a03-cd888dd1249b" containerName="controller-manager" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.918540 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="13a78439-0d99-4068-8a03-cd888dd1249b" containerName="controller-manager" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.918661 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="13a78439-0d99-4068-8a03-cd888dd1249b" containerName="controller-manager" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.919079 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:09 crc kubenswrapper[4725]: I1202 13:10:09.968989 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l"] Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.057744 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-proxy-ca-bundles\") pod \"13a78439-0d99-4068-8a03-cd888dd1249b\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.057794 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7p49\" (UniqueName: \"kubernetes.io/projected/13a78439-0d99-4068-8a03-cd888dd1249b-kube-api-access-v7p49\") pod \"13a78439-0d99-4068-8a03-cd888dd1249b\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.057841 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13a78439-0d99-4068-8a03-cd888dd1249b-serving-cert\") pod \"13a78439-0d99-4068-8a03-cd888dd1249b\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.057891 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-config\") pod \"13a78439-0d99-4068-8a03-cd888dd1249b\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.057918 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-client-ca\") pod \"13a78439-0d99-4068-8a03-cd888dd1249b\" (UID: \"13a78439-0d99-4068-8a03-cd888dd1249b\") " Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.057996 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-client-ca\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.058037 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/befb7e08-5919-4fda-b0fd-e25ff4970f5d-serving-cert\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.058056 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-config\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.058131 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-proxy-ca-bundles\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.058268 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn8vm\" (UniqueName: \"kubernetes.io/projected/befb7e08-5919-4fda-b0fd-e25ff4970f5d-kube-api-access-jn8vm\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.058583 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-client-ca" (OuterVolumeSpecName: "client-ca") pod "13a78439-0d99-4068-8a03-cd888dd1249b" (UID: "13a78439-0d99-4068-8a03-cd888dd1249b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.058638 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-config" (OuterVolumeSpecName: "config") pod "13a78439-0d99-4068-8a03-cd888dd1249b" (UID: "13a78439-0d99-4068-8a03-cd888dd1249b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.059048 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "13a78439-0d99-4068-8a03-cd888dd1249b" (UID: "13a78439-0d99-4068-8a03-cd888dd1249b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.075079 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13a78439-0d99-4068-8a03-cd888dd1249b-kube-api-access-v7p49" (OuterVolumeSpecName: "kube-api-access-v7p49") pod "13a78439-0d99-4068-8a03-cd888dd1249b" (UID: "13a78439-0d99-4068-8a03-cd888dd1249b"). InnerVolumeSpecName "kube-api-access-v7p49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.075691 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13a78439-0d99-4068-8a03-cd888dd1249b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "13a78439-0d99-4068-8a03-cd888dd1249b" (UID: "13a78439-0d99-4068-8a03-cd888dd1249b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159681 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-proxy-ca-bundles\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159732 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn8vm\" (UniqueName: \"kubernetes.io/projected/befb7e08-5919-4fda-b0fd-e25ff4970f5d-kube-api-access-jn8vm\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159788 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-client-ca\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159814 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/befb7e08-5919-4fda-b0fd-e25ff4970f5d-serving-cert\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159835 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-config\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159869 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159880 4725 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159889 4725 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/13a78439-0d99-4068-8a03-cd888dd1249b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159898 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7p49\" (UniqueName: \"kubernetes.io/projected/13a78439-0d99-4068-8a03-cd888dd1249b-kube-api-access-v7p49\") on node \"crc\" DevicePath \"\"" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.159907 4725 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13a78439-0d99-4068-8a03-cd888dd1249b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.160786 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-client-ca\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.161009 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-config\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.161168 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/befb7e08-5919-4fda-b0fd-e25ff4970f5d-proxy-ca-bundles\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.164177 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/befb7e08-5919-4fda-b0fd-e25ff4970f5d-serving-cert\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.173750 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn8vm\" (UniqueName: \"kubernetes.io/projected/befb7e08-5919-4fda-b0fd-e25ff4970f5d-kube-api-access-jn8vm\") pod \"controller-manager-7cd8b9fcfc-q5s7l\" (UID: \"befb7e08-5919-4fda-b0fd-e25ff4970f5d\") " pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.245098 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.601294 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.612611 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f6fc7c544-fcs86" event={"ID":"13a78439-0d99-4068-8a03-cd888dd1249b","Type":"ContainerDied","Data":"8f2be59e927d6e25b99b5559d2ac9224fd183037378cd1c6dd3ae750fa503487"} Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.612669 4725 scope.go:117] "RemoveContainer" containerID="170afded2c6443277e7d82a79804cea82fce5fd8f79fe3b6481a2e611fecffac" Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.636713 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f6fc7c544-fcs86"] Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.641319 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6f6fc7c544-fcs86"] Dec 02 13:10:10 crc kubenswrapper[4725]: I1202 13:10:10.668298 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l"] Dec 02 13:10:10 crc kubenswrapper[4725]: W1202 13:10:10.674898 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbefb7e08_5919_4fda_b0fd_e25ff4970f5d.slice/crio-53a9d3ee5dace96af84e0cdefdff755c8ad33d1f41029d90beeadd532a412e23 WatchSource:0}: Error finding container 53a9d3ee5dace96af84e0cdefdff755c8ad33d1f41029d90beeadd532a412e23: Status 404 returned error can't find the container with id 53a9d3ee5dace96af84e0cdefdff755c8ad33d1f41029d90beeadd532a412e23 Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.274222 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13a78439-0d99-4068-8a03-cd888dd1249b" path="/var/lib/kubelet/pods/13a78439-0d99-4068-8a03-cd888dd1249b/volumes" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.607561 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" event={"ID":"befb7e08-5919-4fda-b0fd-e25ff4970f5d","Type":"ContainerStarted","Data":"bea2befee8a1a53279f744735fb168e598c2aab5ea09802c744c699e773ae0a2"} Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.607605 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" event={"ID":"befb7e08-5919-4fda-b0fd-e25ff4970f5d","Type":"ContainerStarted","Data":"53a9d3ee5dace96af84e0cdefdff755c8ad33d1f41029d90beeadd532a412e23"} Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.607967 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.609166 4725 patch_prober.go:28] interesting pod/controller-manager-7cd8b9fcfc-q5s7l container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.609206 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" podUID="befb7e08-5919-4fda-b0fd-e25ff4970f5d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.631572 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" podStartSLOduration=3.631535909 podStartE2EDuration="3.631535909s" podCreationTimestamp="2025-12-02 13:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:10:11.626234547 +0000 UTC m=+342.582876242" watchObservedRunningTime="2025-12-02 13:10:11.631535909 +0000 UTC m=+342.588177604" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.822300 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.822370 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.859950 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.969050 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:11 crc kubenswrapper[4725]: I1202 13:10:11.969089 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:12 crc kubenswrapper[4725]: I1202 13:10:12.006104 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:12 crc kubenswrapper[4725]: I1202 13:10:12.616077 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" Dec 02 13:10:12 crc kubenswrapper[4725]: I1202 13:10:12.663764 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:10:12 crc kubenswrapper[4725]: I1202 13:10:12.669800 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rs8r5" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.740958 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r6bw4"] Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.741733 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.753277 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r6bw4"] Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.928713 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-registry-tls\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.928769 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w58k2\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-kube-api-access-w58k2\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.928791 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a265a6d-553e-4db3-a5c9-7c403a079078-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.928962 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a265a6d-553e-4db3-a5c9-7c403a079078-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.929011 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a265a6d-553e-4db3-a5c9-7c403a079078-registry-certificates\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.929037 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.929130 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a265a6d-553e-4db3-a5c9-7c403a079078-trusted-ca\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.929221 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-bound-sa-token\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:15 crc kubenswrapper[4725]: I1202 13:10:15.957336 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.030913 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-registry-tls\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.030970 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w58k2\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-kube-api-access-w58k2\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.030986 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a265a6d-553e-4db3-a5c9-7c403a079078-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.031023 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a265a6d-553e-4db3-a5c9-7c403a079078-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.031043 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a265a6d-553e-4db3-a5c9-7c403a079078-registry-certificates\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.031060 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a265a6d-553e-4db3-a5c9-7c403a079078-trusted-ca\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.031083 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-bound-sa-token\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.031535 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a265a6d-553e-4db3-a5c9-7c403a079078-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.032184 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a265a6d-553e-4db3-a5c9-7c403a079078-registry-certificates\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.032276 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a265a6d-553e-4db3-a5c9-7c403a079078-trusted-ca\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.037008 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-registry-tls\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.047302 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a265a6d-553e-4db3-a5c9-7c403a079078-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.051005 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w58k2\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-kube-api-access-w58k2\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.051220 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a265a6d-553e-4db3-a5c9-7c403a079078-bound-sa-token\") pod \"image-registry-66df7c8f76-r6bw4\" (UID: \"5a265a6d-553e-4db3-a5c9-7c403a079078\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.067991 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.477614 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r6bw4"] Dec 02 13:10:16 crc kubenswrapper[4725]: I1202 13:10:16.633777 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" event={"ID":"5a265a6d-553e-4db3-a5c9-7c403a079078","Type":"ContainerStarted","Data":"f24f87cacf788809035abcce4b8cf06dc07bff39a7fb5d779d063dc66601412f"} Dec 02 13:10:17 crc kubenswrapper[4725]: I1202 13:10:17.640981 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" event={"ID":"5a265a6d-553e-4db3-a5c9-7c403a079078","Type":"ContainerStarted","Data":"39c4e548e7d5199fb9ca70ab05cbbfaf1f4551f644b694063e3b5f922312e1e3"} Dec 02 13:10:17 crc kubenswrapper[4725]: I1202 13:10:17.641436 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:17 crc kubenswrapper[4725]: I1202 13:10:17.664316 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" podStartSLOduration=2.664300821 podStartE2EDuration="2.664300821s" podCreationTimestamp="2025-12-02 13:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:10:17.660010925 +0000 UTC m=+348.616652620" watchObservedRunningTime="2025-12-02 13:10:17.664300821 +0000 UTC m=+348.620942516" Dec 02 13:10:24 crc kubenswrapper[4725]: I1202 13:10:24.284929 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:10:24 crc kubenswrapper[4725]: I1202 13:10:24.285283 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:10:36 crc kubenswrapper[4725]: I1202 13:10:36.075833 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-r6bw4" Dec 02 13:10:36 crc kubenswrapper[4725]: I1202 13:10:36.185655 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5m8fq"] Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.285063 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.285613 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.285661 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.288852 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e34e8f201b66e3cbc21d632890c9a1c51e534112ab1f9ab4f4f771f7eb972cdd"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.289068 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://e34e8f201b66e3cbc21d632890c9a1c51e534112ab1f9ab4f4f771f7eb972cdd" gracePeriod=600 Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.838152 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="e34e8f201b66e3cbc21d632890c9a1c51e534112ab1f9ab4f4f771f7eb972cdd" exitCode=0 Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.838235 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"e34e8f201b66e3cbc21d632890c9a1c51e534112ab1f9ab4f4f771f7eb972cdd"} Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.838945 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"7a038ec0fa554f0a9752863aee67be397498e9cc4ec37228c8b10ecc3fbb0599"} Dec 02 13:10:54 crc kubenswrapper[4725]: I1202 13:10:54.838967 4725 scope.go:117] "RemoveContainer" containerID="e8ab098c122ce6deae8692d692b93345ee37e8b1d1e0721acd1e8f00513aac9f" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.221001 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" podUID="7828c47b-d717-48f4-ba0a-375358ec3d79" containerName="registry" containerID="cri-o://78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b" gracePeriod=30 Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.597372 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.733694 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-trusted-ca\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.733780 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7828c47b-d717-48f4-ba0a-375358ec3d79-ca-trust-extracted\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.733841 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmrpd\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-kube-api-access-fmrpd\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.733868 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-tls\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.733905 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-certificates\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.734064 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.734091 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-bound-sa-token\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.734120 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7828c47b-d717-48f4-ba0a-375358ec3d79-installation-pull-secrets\") pod \"7828c47b-d717-48f4-ba0a-375358ec3d79\" (UID: \"7828c47b-d717-48f4-ba0a-375358ec3d79\") " Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.734974 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.734984 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.740286 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7828c47b-d717-48f4-ba0a-375358ec3d79-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.740372 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.740757 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.741306 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-kube-api-access-fmrpd" (OuterVolumeSpecName: "kube-api-access-fmrpd") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "kube-api-access-fmrpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.743890 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.753772 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7828c47b-d717-48f4-ba0a-375358ec3d79-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "7828c47b-d717-48f4-ba0a-375358ec3d79" (UID: "7828c47b-d717-48f4-ba0a-375358ec3d79"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835513 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmrpd\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-kube-api-access-fmrpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835559 4725 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835574 4725 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835585 4725 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7828c47b-d717-48f4-ba0a-375358ec3d79-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835597 4725 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7828c47b-d717-48f4-ba0a-375358ec3d79-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835608 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7828c47b-d717-48f4-ba0a-375358ec3d79-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.835619 4725 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7828c47b-d717-48f4-ba0a-375358ec3d79-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.877961 4725 generic.go:334] "Generic (PLEG): container finished" podID="7828c47b-d717-48f4-ba0a-375358ec3d79" containerID="78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b" exitCode=0 Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.878008 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.878017 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" event={"ID":"7828c47b-d717-48f4-ba0a-375358ec3d79","Type":"ContainerDied","Data":"78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b"} Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.878156 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5m8fq" event={"ID":"7828c47b-d717-48f4-ba0a-375358ec3d79","Type":"ContainerDied","Data":"fe5338e8df5c062e42ccff03660645be848f6dcaaa5a0e594c2c318435bc0229"} Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.878188 4725 scope.go:117] "RemoveContainer" containerID="78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.895527 4725 scope.go:117] "RemoveContainer" containerID="78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b" Dec 02 13:11:01 crc kubenswrapper[4725]: E1202 13:11:01.895971 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b\": container with ID starting with 78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b not found: ID does not exist" containerID="78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.896027 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b"} err="failed to get container status \"78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b\": rpc error: code = NotFound desc = could not find container \"78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b\": container with ID starting with 78e79b2ee64cdb726bc17eb627eb1feca0ad892e4777006ca693e6fbaf9edb8b not found: ID does not exist" Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.905173 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5m8fq"] Dec 02 13:11:01 crc kubenswrapper[4725]: I1202 13:11:01.909644 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5m8fq"] Dec 02 13:11:03 crc kubenswrapper[4725]: I1202 13:11:03.280570 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7828c47b-d717-48f4-ba0a-375358ec3d79" path="/var/lib/kubelet/pods/7828c47b-d717-48f4-ba0a-375358ec3d79/volumes" Dec 02 13:12:54 crc kubenswrapper[4725]: I1202 13:12:54.284575 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:12:54 crc kubenswrapper[4725]: I1202 13:12:54.285274 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:13:24 crc kubenswrapper[4725]: I1202 13:13:24.284880 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:13:24 crc kubenswrapper[4725]: I1202 13:13:24.285405 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.284711 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.285158 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.285195 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.285570 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a038ec0fa554f0a9752863aee67be397498e9cc4ec37228c8b10ecc3fbb0599"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.285626 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://7a038ec0fa554f0a9752863aee67be397498e9cc4ec37228c8b10ecc3fbb0599" gracePeriod=600 Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.911104 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="7a038ec0fa554f0a9752863aee67be397498e9cc4ec37228c8b10ecc3fbb0599" exitCode=0 Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.911180 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"7a038ec0fa554f0a9752863aee67be397498e9cc4ec37228c8b10ecc3fbb0599"} Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.911420 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"01e784ee70283b45e319c7d0015626dcb847bd5037e05305c113f2ede968d762"} Dec 02 13:13:54 crc kubenswrapper[4725]: I1202 13:13:54.911445 4725 scope.go:117] "RemoveContainer" containerID="e34e8f201b66e3cbc21d632890c9a1c51e534112ab1f9ab4f4f771f7eb972cdd" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.542353 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f864q"] Dec 02 13:14:30 crc kubenswrapper[4725]: E1202 13:14:30.543057 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7828c47b-d717-48f4-ba0a-375358ec3d79" containerName="registry" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.543068 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7828c47b-d717-48f4-ba0a-375358ec3d79" containerName="registry" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.543160 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7828c47b-d717-48f4-ba0a-375358ec3d79" containerName="registry" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.543539 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.544807 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.545901 4725 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-mhm8f" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.546526 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.553022 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f864q"] Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.573841 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hsrpm"] Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.574686 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.576623 4725 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qf4lp" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.579713 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6prlb"] Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.580492 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6prlb" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.584678 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hsrpm"] Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.586431 4725 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-28n8x" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.588771 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6prlb"] Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.644309 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt6th\" (UniqueName: \"kubernetes.io/projected/f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4-kube-api-access-vt6th\") pod \"cert-manager-webhook-5655c58dd6-hsrpm\" (UID: \"f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.644553 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m5rf\" (UniqueName: \"kubernetes.io/projected/864d4bd0-504a-4d91-bb28-b920fe8845aa-kube-api-access-4m5rf\") pod \"cert-manager-cainjector-7f985d654d-f864q\" (UID: \"864d4bd0-504a-4d91-bb28-b920fe8845aa\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.644574 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvf8k\" (UniqueName: \"kubernetes.io/projected/81141996-2325-4372-81be-7f69cbec7194-kube-api-access-bvf8k\") pod \"cert-manager-5b446d88c5-6prlb\" (UID: \"81141996-2325-4372-81be-7f69cbec7194\") " pod="cert-manager/cert-manager-5b446d88c5-6prlb" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.745175 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m5rf\" (UniqueName: \"kubernetes.io/projected/864d4bd0-504a-4d91-bb28-b920fe8845aa-kube-api-access-4m5rf\") pod \"cert-manager-cainjector-7f985d654d-f864q\" (UID: \"864d4bd0-504a-4d91-bb28-b920fe8845aa\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.745212 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvf8k\" (UniqueName: \"kubernetes.io/projected/81141996-2325-4372-81be-7f69cbec7194-kube-api-access-bvf8k\") pod \"cert-manager-5b446d88c5-6prlb\" (UID: \"81141996-2325-4372-81be-7f69cbec7194\") " pod="cert-manager/cert-manager-5b446d88c5-6prlb" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.745245 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt6th\" (UniqueName: \"kubernetes.io/projected/f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4-kube-api-access-vt6th\") pod \"cert-manager-webhook-5655c58dd6-hsrpm\" (UID: \"f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.763046 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvf8k\" (UniqueName: \"kubernetes.io/projected/81141996-2325-4372-81be-7f69cbec7194-kube-api-access-bvf8k\") pod \"cert-manager-5b446d88c5-6prlb\" (UID: \"81141996-2325-4372-81be-7f69cbec7194\") " pod="cert-manager/cert-manager-5b446d88c5-6prlb" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.763087 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt6th\" (UniqueName: \"kubernetes.io/projected/f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4-kube-api-access-vt6th\") pod \"cert-manager-webhook-5655c58dd6-hsrpm\" (UID: \"f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.763204 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m5rf\" (UniqueName: \"kubernetes.io/projected/864d4bd0-504a-4d91-bb28-b920fe8845aa-kube-api-access-4m5rf\") pod \"cert-manager-cainjector-7f985d654d-f864q\" (UID: \"864d4bd0-504a-4d91-bb28-b920fe8845aa\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.860127 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.891115 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:30 crc kubenswrapper[4725]: I1202 13:14:30.899323 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6prlb" Dec 02 13:14:31 crc kubenswrapper[4725]: I1202 13:14:31.092097 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-f864q"] Dec 02 13:14:31 crc kubenswrapper[4725]: I1202 13:14:31.100062 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:14:31 crc kubenswrapper[4725]: I1202 13:14:31.326566 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-hsrpm"] Dec 02 13:14:31 crc kubenswrapper[4725]: W1202 13:14:31.329496 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf84fbbc8_bcb0_4aff_b5a4_48e932ec4ac4.slice/crio-949cddaf0501a18214a4b927b0239f858fe7827f990308a030892adae5a4f259 WatchSource:0}: Error finding container 949cddaf0501a18214a4b927b0239f858fe7827f990308a030892adae5a4f259: Status 404 returned error can't find the container with id 949cddaf0501a18214a4b927b0239f858fe7827f990308a030892adae5a4f259 Dec 02 13:14:31 crc kubenswrapper[4725]: I1202 13:14:31.364764 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6prlb"] Dec 02 13:14:32 crc kubenswrapper[4725]: I1202 13:14:32.103673 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" event={"ID":"864d4bd0-504a-4d91-bb28-b920fe8845aa","Type":"ContainerStarted","Data":"5b4b08a356d8871dc7fcfb3cae2b746d281eaa339acf414f340a6fa4baef2059"} Dec 02 13:14:32 crc kubenswrapper[4725]: I1202 13:14:32.105192 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6prlb" event={"ID":"81141996-2325-4372-81be-7f69cbec7194","Type":"ContainerStarted","Data":"dd492f21f9e3501a4b246fafe39dcf96874c2a869083a3e582a92ed5879332ae"} Dec 02 13:14:32 crc kubenswrapper[4725]: I1202 13:14:32.106427 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" event={"ID":"f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4","Type":"ContainerStarted","Data":"949cddaf0501a18214a4b927b0239f858fe7827f990308a030892adae5a4f259"} Dec 02 13:14:34 crc kubenswrapper[4725]: I1202 13:14:34.123025 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6prlb" event={"ID":"81141996-2325-4372-81be-7f69cbec7194","Type":"ContainerStarted","Data":"a41f1f1b5721a3f7cf35af85466dd6ea7283a55eda00a276fdf95a6ae9d86a48"} Dec 02 13:14:34 crc kubenswrapper[4725]: I1202 13:14:34.128447 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" event={"ID":"f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4","Type":"ContainerStarted","Data":"c949b900d3ab0adcb8cf7b36c219d5657b334d5eed096fdd818bf56c30f84b37"} Dec 02 13:14:34 crc kubenswrapper[4725]: I1202 13:14:34.128786 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:34 crc kubenswrapper[4725]: I1202 13:14:34.149247 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-6prlb" podStartSLOduration=1.702290292 podStartE2EDuration="4.149210289s" podCreationTimestamp="2025-12-02 13:14:30 +0000 UTC" firstStartedPulling="2025-12-02 13:14:31.371448419 +0000 UTC m=+602.328090114" lastFinishedPulling="2025-12-02 13:14:33.818368416 +0000 UTC m=+604.775010111" observedRunningTime="2025-12-02 13:14:34.138434478 +0000 UTC m=+605.095076173" watchObservedRunningTime="2025-12-02 13:14:34.149210289 +0000 UTC m=+605.105851984" Dec 02 13:14:34 crc kubenswrapper[4725]: I1202 13:14:34.162849 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" podStartSLOduration=1.7269622629999999 podStartE2EDuration="4.162830772s" podCreationTimestamp="2025-12-02 13:14:30 +0000 UTC" firstStartedPulling="2025-12-02 13:14:31.332225872 +0000 UTC m=+602.288867567" lastFinishedPulling="2025-12-02 13:14:33.768094381 +0000 UTC m=+604.724736076" observedRunningTime="2025-12-02 13:14:34.156942404 +0000 UTC m=+605.113584099" watchObservedRunningTime="2025-12-02 13:14:34.162830772 +0000 UTC m=+605.119472467" Dec 02 13:14:35 crc kubenswrapper[4725]: I1202 13:14:35.133768 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" event={"ID":"864d4bd0-504a-4d91-bb28-b920fe8845aa","Type":"ContainerStarted","Data":"09ccc61a2c114ce5e79b2dd4793e3da4b8e69e3e82f29debba077f449e5a6871"} Dec 02 13:14:35 crc kubenswrapper[4725]: I1202 13:14:35.149954 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-f864q" podStartSLOduration=1.8663739019999999 podStartE2EDuration="5.149934458s" podCreationTimestamp="2025-12-02 13:14:30 +0000 UTC" firstStartedPulling="2025-12-02 13:14:31.099836395 +0000 UTC m=+602.056478090" lastFinishedPulling="2025-12-02 13:14:34.383396951 +0000 UTC m=+605.340038646" observedRunningTime="2025-12-02 13:14:35.145169608 +0000 UTC m=+606.101811303" watchObservedRunningTime="2025-12-02 13:14:35.149934458 +0000 UTC m=+606.106576153" Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.796983 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nnqsk"] Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.799339 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="sbdb" containerID="cri-o://4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.799555 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="nbdb" containerID="cri-o://91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.799661 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="northd" containerID="cri-o://bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.798337 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-controller" containerID="cri-o://74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.799796 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.799950 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-node" containerID="cri-o://6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.799928 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-acl-logging" containerID="cri-o://2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.832624 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" containerID="cri-o://5f43d87f4c9f0d0bf3882ecef178e296d0b9d78e650fcf8243c0af4028c4a8b6" gracePeriod=30 Dec 02 13:14:40 crc kubenswrapper[4725]: I1202 13:14:40.894237 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-hsrpm" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.165316 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/2.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.165810 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/1.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.165856 4725 generic.go:334] "Generic (PLEG): container finished" podID="4a8c02a6-36ad-4a9c-88b0-064dcc9b7327" containerID="d654f554b44b0403cae91f5d8143c150d9428226b7a282a7666f50eab5f10622" exitCode=2 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.165919 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerDied","Data":"d654f554b44b0403cae91f5d8143c150d9428226b7a282a7666f50eab5f10622"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.165957 4725 scope.go:117] "RemoveContainer" containerID="9e585c49bf99faf9bde400ed17376dfddfd065ba4125b12e992db4e7fec77913" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.168443 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovnkube-controller/3.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.168488 4725 scope.go:117] "RemoveContainer" containerID="d654f554b44b0403cae91f5d8143c150d9428226b7a282a7666f50eab5f10622" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.169006 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8s8qq_openshift-multus(4a8c02a6-36ad-4a9c-88b0-064dcc9b7327)\"" pod="openshift-multus/multus-8s8qq" podUID="4a8c02a6-36ad-4a9c-88b0-064dcc9b7327" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.182160 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovn-acl-logging/0.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.184295 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovn-controller/0.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188160 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="5f43d87f4c9f0d0bf3882ecef178e296d0b9d78e650fcf8243c0af4028c4a8b6" exitCode=0 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188200 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8" exitCode=0 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188211 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506" exitCode=0 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188220 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2" exitCode=0 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188229 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a" exitCode=0 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188238 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e" exitCode=0 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188249 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668" exitCode=143 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188260 4725 generic.go:334] "Generic (PLEG): container finished" podID="f939ed3f-9402-4a57-858f-0323084742a9" containerID="74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07" exitCode=143 Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188288 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"5f43d87f4c9f0d0bf3882ecef178e296d0b9d78e650fcf8243c0af4028c4a8b6"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188325 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188340 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188353 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188365 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188376 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188387 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.188398 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07"} Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.401836 4725 scope.go:117] "RemoveContainer" containerID="acb3a76b2456e4cf34fe8f76dda72592878c300f086d3323eb16e6db5d521ae0" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.566350 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovn-acl-logging/0.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.567002 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovn-controller/0.log" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.567394 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.615732 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pftpf"] Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616261 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="nbdb" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616285 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="nbdb" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616299 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-node" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616307 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-node" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616339 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616375 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616411 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="sbdb" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616418 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="sbdb" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616427 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616433 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616442 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616448 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616541 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kubecfg-setup" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616645 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kubecfg-setup" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616655 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="northd" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616660 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="northd" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616668 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616675 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616682 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616689 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616700 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-acl-logging" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616706 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-acl-logging" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.616717 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616725 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616840 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="sbdb" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616852 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616861 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616869 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616879 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-node" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616888 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="nbdb" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616897 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616906 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovn-acl-logging" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616914 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.616922 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="northd" Dec 02 13:14:41 crc kubenswrapper[4725]: E1202 13:14:41.617028 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.617039 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.617150 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.617164 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f939ed3f-9402-4a57-858f-0323084742a9" containerName="ovnkube-controller" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.619249 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679651 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btq9p\" (UniqueName: \"kubernetes.io/projected/f939ed3f-9402-4a57-858f-0323084742a9-kube-api-access-btq9p\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679724 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-systemd\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679747 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-etc-openvswitch\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679789 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679814 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-env-overrides\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679836 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-netns\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679851 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-kubelet\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679875 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-var-lib-openvswitch\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.679923 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680061 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-openvswitch\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680093 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680468 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680528 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680552 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680583 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680633 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-ovn\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680660 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680671 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680770 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-bin\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680801 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f939ed3f-9402-4a57-858f-0323084742a9-ovn-node-metrics-cert\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680844 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-slash\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680867 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-node-log\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680888 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-script-lib\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680901 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-slash" (OuterVolumeSpecName: "host-slash") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680909 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-log-socket\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680934 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-log-socket" (OuterVolumeSpecName: "log-socket") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680960 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-node-log" (OuterVolumeSpecName: "node-log") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.680994 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-netd\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681034 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-config\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681077 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-systemd-units\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681101 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-ovn-kubernetes\") pod \"f939ed3f-9402-4a57-858f-0323084742a9\" (UID: \"f939ed3f-9402-4a57-858f-0323084742a9\") " Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681204 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-systemd-units\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681268 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-ovn\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681294 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovnkube-script-lib\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681348 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-cni-bin\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681372 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovnkube-config\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681390 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-slash\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681441 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681463 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681507 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovn-node-metrics-cert\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681594 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-etc-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681609 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681619 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681778 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-var-lib-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681821 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-kubelet\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681845 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-cni-netd\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681870 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.681909 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682074 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-systemd\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682159 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm22r\" (UniqueName: \"kubernetes.io/projected/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-kube-api-access-jm22r\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682081 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682312 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-node-log\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682329 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682361 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682383 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-log-socket\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682427 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-run-netns\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682531 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-env-overrides\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682655 4725 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682676 4725 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682687 4725 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682699 4725 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682718 4725 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682736 4725 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682746 4725 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682759 4725 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682771 4725 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682781 4725 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682792 4725 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682801 4725 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682809 4725 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682820 4725 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-host-slash\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682828 4725 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-node-log\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682837 4725 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f939ed3f-9402-4a57-858f-0323084742a9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.682845 4725 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-log-socket\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.685398 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f939ed3f-9402-4a57-858f-0323084742a9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.685451 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f939ed3f-9402-4a57-858f-0323084742a9-kube-api-access-btq9p" (OuterVolumeSpecName: "kube-api-access-btq9p") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "kube-api-access-btq9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.693073 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "f939ed3f-9402-4a57-858f-0323084742a9" (UID: "f939ed3f-9402-4a57-858f-0323084742a9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784267 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-env-overrides\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784329 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-systemd-units\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784356 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-ovn\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784382 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovnkube-script-lib\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784404 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-cni-bin\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784455 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovnkube-config\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784511 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-slash\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784507 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-systemd-units\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784536 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovn-node-metrics-cert\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784554 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-etc-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784576 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784587 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-cni-bin\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784599 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-var-lib-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784620 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-kubelet\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784625 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-ovn\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784637 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-cni-netd\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784656 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784687 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-systemd\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784703 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm22r\" (UniqueName: \"kubernetes.io/projected/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-kube-api-access-jm22r\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784724 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-node-log\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784742 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784756 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-log-socket\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784771 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-run-netns\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784831 4725 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f939ed3f-9402-4a57-858f-0323084742a9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784843 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btq9p\" (UniqueName: \"kubernetes.io/projected/f939ed3f-9402-4a57-858f-0323084742a9-kube-api-access-btq9p\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784854 4725 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f939ed3f-9402-4a57-858f-0323084742a9-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784865 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-kubelet\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784894 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-run-netns\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784917 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-etc-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784931 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-cni-netd\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784957 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784962 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.784977 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-systemd\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.785002 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-var-lib-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.785037 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-run-openvswitch\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.785066 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-node-log\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.785097 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-log-socket\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.785322 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-host-slash\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.785678 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-env-overrides\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.786053 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovnkube-script-lib\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.786570 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovnkube-config\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.788378 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-ovn-node-metrics-cert\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.800863 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm22r\" (UniqueName: \"kubernetes.io/projected/8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd-kube-api-access-jm22r\") pod \"ovnkube-node-pftpf\" (UID: \"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:41 crc kubenswrapper[4725]: I1202 13:14:41.944040 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.193559 4725 generic.go:334] "Generic (PLEG): container finished" podID="8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd" containerID="0138da0bd54567f1bd732b3d65c3fb705780d3ca4eff6a19920de1d088d4efb2" exitCode=0 Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.193617 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerDied","Data":"0138da0bd54567f1bd732b3d65c3fb705780d3ca4eff6a19920de1d088d4efb2"} Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.193647 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"55613e131c997da93b5e0530f24061fb81b0273500d69b8fd546009a023ef9f5"} Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.196896 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/2.log" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.201272 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovn-acl-logging/0.log" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.202371 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nnqsk_f939ed3f-9402-4a57-858f-0323084742a9/ovn-controller/0.log" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.203005 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" event={"ID":"f939ed3f-9402-4a57-858f-0323084742a9","Type":"ContainerDied","Data":"8d7ef5e525ac20549b36ea374fb347ef713cde9d74c4c74273f293bc146a06be"} Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.203054 4725 scope.go:117] "RemoveContainer" containerID="5f43d87f4c9f0d0bf3882ecef178e296d0b9d78e650fcf8243c0af4028c4a8b6" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.203152 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nnqsk" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.224779 4725 scope.go:117] "RemoveContainer" containerID="4f929b105487a21ae619ce37e455ceb389b27cb57e084927e928d1c62fc670a8" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.254751 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nnqsk"] Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.259112 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nnqsk"] Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.262903 4725 scope.go:117] "RemoveContainer" containerID="91f0d09bd516ad9116f0d4c3b9237d13b23f1c82e04dbea3aaf114bad0ec5506" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.275284 4725 scope.go:117] "RemoveContainer" containerID="bbcdb00c64ab41db9a67a8279083b92442bd8057cff13f30cebfe7963c8980f2" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.297839 4725 scope.go:117] "RemoveContainer" containerID="9be521d64a77325b0ea9770b43f2c72164897d0a112d555a4107f279761e989a" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.313321 4725 scope.go:117] "RemoveContainer" containerID="6c677c90aa91cbf5b483d4f1a87a88416d011e512d224569c79e771cbb8a6c1e" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.327524 4725 scope.go:117] "RemoveContainer" containerID="2ce2b328a2b475d948ffe8ce257594def6b8bc3387024f0369dbf2718093e668" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.344239 4725 scope.go:117] "RemoveContainer" containerID="74f5e02943d64adfcdbb3f3136d71f0abc4ed863ae02ac526d1e15c8cb43ff07" Dec 02 13:14:42 crc kubenswrapper[4725]: I1202 13:14:42.390706 4725 scope.go:117] "RemoveContainer" containerID="143bbcbb482091ab96527c889265fd9a114aafb98f0df48b6a970930c6d992e2" Dec 02 13:14:43 crc kubenswrapper[4725]: I1202 13:14:43.222168 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"a5cf947817e23b043aad659baf0e52fb0252c835befded4e123bf33e6723620e"} Dec 02 13:14:43 crc kubenswrapper[4725]: I1202 13:14:43.222211 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"2203caabcc7e1502702736c3c4e4af399a98779edc295f85ffcf1ed83cf9d7ea"} Dec 02 13:14:43 crc kubenswrapper[4725]: I1202 13:14:43.222221 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"7d907def82bca8246793af2d0dc8cdea7e2f906e7da128be69709e3341aa331c"} Dec 02 13:14:43 crc kubenswrapper[4725]: I1202 13:14:43.222231 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"f4fcc7776643486c75cfe60b6637c4051f22fad734c079f573d01b88ddbf2294"} Dec 02 13:14:43 crc kubenswrapper[4725]: I1202 13:14:43.222243 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"af05e75a793c33bbc24c0e8a1a7a7a948efa85d6963e5792a0b528f315e42492"} Dec 02 13:14:43 crc kubenswrapper[4725]: I1202 13:14:43.274978 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f939ed3f-9402-4a57-858f-0323084742a9" path="/var/lib/kubelet/pods/f939ed3f-9402-4a57-858f-0323084742a9/volumes" Dec 02 13:14:44 crc kubenswrapper[4725]: I1202 13:14:44.230356 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"b7f0b38c5043b3ef7f4e884be80ed943c3ce0e350701c72e312dd123356c5cc0"} Dec 02 13:14:46 crc kubenswrapper[4725]: I1202 13:14:46.242259 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"60c97fe630308afa96048f510910d4b3ce749a3d835e59de643664d851b95530"} Dec 02 13:14:48 crc kubenswrapper[4725]: I1202 13:14:48.256280 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" event={"ID":"8c7ae0b2-48a3-4755-a2a5-bb6f60f7f3bd","Type":"ContainerStarted","Data":"534e3ac47469d7fd626aef2f25ca0a67bf40f10eca70e63f561c4fc97b5a79ee"} Dec 02 13:14:48 crc kubenswrapper[4725]: I1202 13:14:48.256859 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:48 crc kubenswrapper[4725]: I1202 13:14:48.256874 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:48 crc kubenswrapper[4725]: I1202 13:14:48.281090 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:48 crc kubenswrapper[4725]: I1202 13:14:48.284067 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" podStartSLOduration=7.284054193 podStartE2EDuration="7.284054193s" podCreationTimestamp="2025-12-02 13:14:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:14:48.282672939 +0000 UTC m=+619.239314634" watchObservedRunningTime="2025-12-02 13:14:48.284054193 +0000 UTC m=+619.240695888" Dec 02 13:14:49 crc kubenswrapper[4725]: I1202 13:14:49.262185 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:49 crc kubenswrapper[4725]: I1202 13:14:49.288042 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:14:53 crc kubenswrapper[4725]: I1202 13:14:53.268936 4725 scope.go:117] "RemoveContainer" containerID="d654f554b44b0403cae91f5d8143c150d9428226b7a282a7666f50eab5f10622" Dec 02 13:14:53 crc kubenswrapper[4725]: E1202 13:14:53.270674 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8s8qq_openshift-multus(4a8c02a6-36ad-4a9c-88b0-064dcc9b7327)\"" pod="openshift-multus/multus-8s8qq" podUID="4a8c02a6-36ad-4a9c-88b0-064dcc9b7327" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.167494 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl"] Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.168643 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.170612 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.172709 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.181575 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl"] Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.341050 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b80abbd6-c57b-484e-9d79-c0809affe488-config-volume\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.341134 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b80abbd6-c57b-484e-9d79-c0809affe488-secret-volume\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.341192 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-972jf\" (UniqueName: \"kubernetes.io/projected/b80abbd6-c57b-484e-9d79-c0809affe488-kube-api-access-972jf\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.442092 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b80abbd6-c57b-484e-9d79-c0809affe488-secret-volume\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.442177 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-972jf\" (UniqueName: \"kubernetes.io/projected/b80abbd6-c57b-484e-9d79-c0809affe488-kube-api-access-972jf\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.442204 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b80abbd6-c57b-484e-9d79-c0809affe488-config-volume\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.443204 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b80abbd6-c57b-484e-9d79-c0809affe488-config-volume\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.447631 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b80abbd6-c57b-484e-9d79-c0809affe488-secret-volume\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.458609 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-972jf\" (UniqueName: \"kubernetes.io/projected/b80abbd6-c57b-484e-9d79-c0809affe488-kube-api-access-972jf\") pod \"collect-profiles-29411355-vhtcl\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: I1202 13:15:00.492366 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: E1202 13:15:00.520993 4725 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(8bd7fa4848beb1508672c06b46f92e1e67c9344cb0150e408982bca8cd2946c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 13:15:00 crc kubenswrapper[4725]: E1202 13:15:00.521067 4725 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(8bd7fa4848beb1508672c06b46f92e1e67c9344cb0150e408982bca8cd2946c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: E1202 13:15:00.521090 4725 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(8bd7fa4848beb1508672c06b46f92e1e67c9344cb0150e408982bca8cd2946c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:00 crc kubenswrapper[4725]: E1202 13:15:00.521137 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager(b80abbd6-c57b-484e-9d79-c0809affe488)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager(b80abbd6-c57b-484e-9d79-c0809affe488)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(8bd7fa4848beb1508672c06b46f92e1e67c9344cb0150e408982bca8cd2946c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" podUID="b80abbd6-c57b-484e-9d79-c0809affe488" Dec 02 13:15:01 crc kubenswrapper[4725]: I1202 13:15:01.320259 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:01 crc kubenswrapper[4725]: I1202 13:15:01.320751 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:01 crc kubenswrapper[4725]: E1202 13:15:01.338340 4725 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(751c45f7053a3cff86c459990e70b5ef796fed3478433cf25c2298abdab66685): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 13:15:01 crc kubenswrapper[4725]: E1202 13:15:01.338412 4725 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(751c45f7053a3cff86c459990e70b5ef796fed3478433cf25c2298abdab66685): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:01 crc kubenswrapper[4725]: E1202 13:15:01.338443 4725 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(751c45f7053a3cff86c459990e70b5ef796fed3478433cf25c2298abdab66685): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:01 crc kubenswrapper[4725]: E1202 13:15:01.338513 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager(b80abbd6-c57b-484e-9d79-c0809affe488)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager(b80abbd6-c57b-484e-9d79-c0809affe488)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29411355-vhtcl_openshift-operator-lifecycle-manager_b80abbd6-c57b-484e-9d79-c0809affe488_0(751c45f7053a3cff86c459990e70b5ef796fed3478433cf25c2298abdab66685): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" podUID="b80abbd6-c57b-484e-9d79-c0809affe488" Dec 02 13:15:06 crc kubenswrapper[4725]: I1202 13:15:06.268754 4725 scope.go:117] "RemoveContainer" containerID="d654f554b44b0403cae91f5d8143c150d9428226b7a282a7666f50eab5f10622" Dec 02 13:15:07 crc kubenswrapper[4725]: I1202 13:15:07.350668 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8s8qq_4a8c02a6-36ad-4a9c-88b0-064dcc9b7327/kube-multus/2.log" Dec 02 13:15:07 crc kubenswrapper[4725]: I1202 13:15:07.350973 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8s8qq" event={"ID":"4a8c02a6-36ad-4a9c-88b0-064dcc9b7327","Type":"ContainerStarted","Data":"65657ee7a0266c8894176fc624827c0bcbadc29c18b2b75113f5095841634bdd"} Dec 02 13:15:11 crc kubenswrapper[4725]: I1202 13:15:11.967308 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pftpf" Dec 02 13:15:12 crc kubenswrapper[4725]: I1202 13:15:12.267842 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:12 crc kubenswrapper[4725]: I1202 13:15:12.268312 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:12 crc kubenswrapper[4725]: I1202 13:15:12.645356 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl"] Dec 02 13:15:13 crc kubenswrapper[4725]: I1202 13:15:13.387764 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" event={"ID":"b80abbd6-c57b-484e-9d79-c0809affe488","Type":"ContainerStarted","Data":"33e6c40f5fd039ed85bfa4f7d396640baa09eae2c4aed6594a7d1d1eb957b7a8"} Dec 02 13:15:13 crc kubenswrapper[4725]: I1202 13:15:13.389560 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" event={"ID":"b80abbd6-c57b-484e-9d79-c0809affe488","Type":"ContainerStarted","Data":"94328defc3542dbd8c6ef267e5bd366c1022fab64270680a0639b6fadb00f61b"} Dec 02 13:15:13 crc kubenswrapper[4725]: I1202 13:15:13.407666 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" podStartSLOduration=13.407637351 podStartE2EDuration="13.407637351s" podCreationTimestamp="2025-12-02 13:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:15:13.402094442 +0000 UTC m=+644.358736147" watchObservedRunningTime="2025-12-02 13:15:13.407637351 +0000 UTC m=+644.364279056" Dec 02 13:15:14 crc kubenswrapper[4725]: I1202 13:15:14.392518 4725 generic.go:334] "Generic (PLEG): container finished" podID="b80abbd6-c57b-484e-9d79-c0809affe488" containerID="33e6c40f5fd039ed85bfa4f7d396640baa09eae2c4aed6594a7d1d1eb957b7a8" exitCode=0 Dec 02 13:15:14 crc kubenswrapper[4725]: I1202 13:15:14.392598 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" event={"ID":"b80abbd6-c57b-484e-9d79-c0809affe488","Type":"ContainerDied","Data":"33e6c40f5fd039ed85bfa4f7d396640baa09eae2c4aed6594a7d1d1eb957b7a8"} Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.628209 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.819655 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-972jf\" (UniqueName: \"kubernetes.io/projected/b80abbd6-c57b-484e-9d79-c0809affe488-kube-api-access-972jf\") pod \"b80abbd6-c57b-484e-9d79-c0809affe488\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.819786 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b80abbd6-c57b-484e-9d79-c0809affe488-secret-volume\") pod \"b80abbd6-c57b-484e-9d79-c0809affe488\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.819872 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b80abbd6-c57b-484e-9d79-c0809affe488-config-volume\") pod \"b80abbd6-c57b-484e-9d79-c0809affe488\" (UID: \"b80abbd6-c57b-484e-9d79-c0809affe488\") " Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.820587 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b80abbd6-c57b-484e-9d79-c0809affe488-config-volume" (OuterVolumeSpecName: "config-volume") pod "b80abbd6-c57b-484e-9d79-c0809affe488" (UID: "b80abbd6-c57b-484e-9d79-c0809affe488"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.825434 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80abbd6-c57b-484e-9d79-c0809affe488-kube-api-access-972jf" (OuterVolumeSpecName: "kube-api-access-972jf") pod "b80abbd6-c57b-484e-9d79-c0809affe488" (UID: "b80abbd6-c57b-484e-9d79-c0809affe488"). InnerVolumeSpecName "kube-api-access-972jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.826778 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80abbd6-c57b-484e-9d79-c0809affe488-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b80abbd6-c57b-484e-9d79-c0809affe488" (UID: "b80abbd6-c57b-484e-9d79-c0809affe488"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.921630 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-972jf\" (UniqueName: \"kubernetes.io/projected/b80abbd6-c57b-484e-9d79-c0809affe488-kube-api-access-972jf\") on node \"crc\" DevicePath \"\"" Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.921669 4725 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b80abbd6-c57b-484e-9d79-c0809affe488-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:15:15 crc kubenswrapper[4725]: I1202 13:15:15.921678 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b80abbd6-c57b-484e-9d79-c0809affe488-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:15:16 crc kubenswrapper[4725]: I1202 13:15:16.406001 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" event={"ID":"b80abbd6-c57b-484e-9d79-c0809affe488","Type":"ContainerDied","Data":"94328defc3542dbd8c6ef267e5bd366c1022fab64270680a0639b6fadb00f61b"} Dec 02 13:15:16 crc kubenswrapper[4725]: I1202 13:15:16.406038 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl" Dec 02 13:15:16 crc kubenswrapper[4725]: I1202 13:15:16.406055 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94328defc3542dbd8c6ef267e5bd366c1022fab64270680a0639b6fadb00f61b" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.738868 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk"] Dec 02 13:15:20 crc kubenswrapper[4725]: E1202 13:15:20.739415 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80abbd6-c57b-484e-9d79-c0809affe488" containerName="collect-profiles" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.739432 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80abbd6-c57b-484e-9d79-c0809affe488" containerName="collect-profiles" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.739594 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80abbd6-c57b-484e-9d79-c0809affe488" containerName="collect-profiles" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.740418 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.742349 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.747778 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk"] Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.786094 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.786141 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn6jn\" (UniqueName: \"kubernetes.io/projected/c0ff45e5-0175-458b-a1e2-a4079690d9b0-kube-api-access-vn6jn\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.786249 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.887965 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.888041 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.888069 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn6jn\" (UniqueName: \"kubernetes.io/projected/c0ff45e5-0175-458b-a1e2-a4079690d9b0-kube-api-access-vn6jn\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.888505 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.888568 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:20 crc kubenswrapper[4725]: I1202 13:15:20.909884 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn6jn\" (UniqueName: \"kubernetes.io/projected/c0ff45e5-0175-458b-a1e2-a4079690d9b0-kube-api-access-vn6jn\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:21 crc kubenswrapper[4725]: I1202 13:15:21.058106 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:21 crc kubenswrapper[4725]: I1202 13:15:21.218773 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk"] Dec 02 13:15:21 crc kubenswrapper[4725]: I1202 13:15:21.430341 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" event={"ID":"c0ff45e5-0175-458b-a1e2-a4079690d9b0","Type":"ContainerStarted","Data":"4ef17b38143a4924d1a28aec2a05f4adb14ab271a468eaf548034761924ec223"} Dec 02 13:15:21 crc kubenswrapper[4725]: I1202 13:15:21.430660 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" event={"ID":"c0ff45e5-0175-458b-a1e2-a4079690d9b0","Type":"ContainerStarted","Data":"9303c8205ca6c4083671507f038d7b1562dbca86d1b179925466eea646bc8069"} Dec 02 13:15:22 crc kubenswrapper[4725]: I1202 13:15:22.437913 4725 generic.go:334] "Generic (PLEG): container finished" podID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerID="4ef17b38143a4924d1a28aec2a05f4adb14ab271a468eaf548034761924ec223" exitCode=0 Dec 02 13:15:22 crc kubenswrapper[4725]: I1202 13:15:22.437960 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" event={"ID":"c0ff45e5-0175-458b-a1e2-a4079690d9b0","Type":"ContainerDied","Data":"4ef17b38143a4924d1a28aec2a05f4adb14ab271a468eaf548034761924ec223"} Dec 02 13:15:24 crc kubenswrapper[4725]: I1202 13:15:24.447895 4725 generic.go:334] "Generic (PLEG): container finished" podID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerID="b55b38759a0fd24df2c1ce3c48d5b41d6e3d07372058568dbd68711ba616fc0a" exitCode=0 Dec 02 13:15:24 crc kubenswrapper[4725]: I1202 13:15:24.448007 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" event={"ID":"c0ff45e5-0175-458b-a1e2-a4079690d9b0","Type":"ContainerDied","Data":"b55b38759a0fd24df2c1ce3c48d5b41d6e3d07372058568dbd68711ba616fc0a"} Dec 02 13:15:25 crc kubenswrapper[4725]: I1202 13:15:25.456490 4725 generic.go:334] "Generic (PLEG): container finished" podID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerID="6e28be3c983c6dd3ec9c1dfa60ebeb797cc8fc4b8a9e66751d31ad7070e9a9c1" exitCode=0 Dec 02 13:15:25 crc kubenswrapper[4725]: I1202 13:15:25.456812 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" event={"ID":"c0ff45e5-0175-458b-a1e2-a4079690d9b0","Type":"ContainerDied","Data":"6e28be3c983c6dd3ec9c1dfa60ebeb797cc8fc4b8a9e66751d31ad7070e9a9c1"} Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.699188 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.877937 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-bundle\") pod \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.878442 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-util\") pod \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.878523 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn6jn\" (UniqueName: \"kubernetes.io/projected/c0ff45e5-0175-458b-a1e2-a4079690d9b0-kube-api-access-vn6jn\") pod \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\" (UID: \"c0ff45e5-0175-458b-a1e2-a4079690d9b0\") " Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.878656 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-bundle" (OuterVolumeSpecName: "bundle") pod "c0ff45e5-0175-458b-a1e2-a4079690d9b0" (UID: "c0ff45e5-0175-458b-a1e2-a4079690d9b0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.878796 4725 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.883695 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ff45e5-0175-458b-a1e2-a4079690d9b0-kube-api-access-vn6jn" (OuterVolumeSpecName: "kube-api-access-vn6jn") pod "c0ff45e5-0175-458b-a1e2-a4079690d9b0" (UID: "c0ff45e5-0175-458b-a1e2-a4079690d9b0"). InnerVolumeSpecName "kube-api-access-vn6jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:15:26 crc kubenswrapper[4725]: I1202 13:15:26.979561 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn6jn\" (UniqueName: \"kubernetes.io/projected/c0ff45e5-0175-458b-a1e2-a4079690d9b0-kube-api-access-vn6jn\") on node \"crc\" DevicePath \"\"" Dec 02 13:15:27 crc kubenswrapper[4725]: I1202 13:15:27.095165 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-util" (OuterVolumeSpecName: "util") pod "c0ff45e5-0175-458b-a1e2-a4079690d9b0" (UID: "c0ff45e5-0175-458b-a1e2-a4079690d9b0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:15:27 crc kubenswrapper[4725]: I1202 13:15:27.181714 4725 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c0ff45e5-0175-458b-a1e2-a4079690d9b0-util\") on node \"crc\" DevicePath \"\"" Dec 02 13:15:27 crc kubenswrapper[4725]: I1202 13:15:27.467014 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" event={"ID":"c0ff45e5-0175-458b-a1e2-a4079690d9b0","Type":"ContainerDied","Data":"9303c8205ca6c4083671507f038d7b1562dbca86d1b179925466eea646bc8069"} Dec 02 13:15:27 crc kubenswrapper[4725]: I1202 13:15:27.467058 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9303c8205ca6c4083671507f038d7b1562dbca86d1b179925466eea646bc8069" Dec 02 13:15:27 crc kubenswrapper[4725]: I1202 13:15:27.467076 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.210096 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq"] Dec 02 13:15:32 crc kubenswrapper[4725]: E1202 13:15:32.210689 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="pull" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.210705 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="pull" Dec 02 13:15:32 crc kubenswrapper[4725]: E1202 13:15:32.210715 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="util" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.210724 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="util" Dec 02 13:15:32 crc kubenswrapper[4725]: E1202 13:15:32.210741 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="extract" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.210749 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="extract" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.210869 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ff45e5-0175-458b-a1e2-a4079690d9b0" containerName="extract" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.211306 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.217174 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-h94p7" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.217225 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.220429 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.225421 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq"] Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.347993 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4zdg\" (UniqueName: \"kubernetes.io/projected/511d65e3-135b-4218-aaeb-bee781e9befb-kube-api-access-r4zdg\") pod \"nmstate-operator-5b5b58f5c8-xz2lq\" (UID: \"511d65e3-135b-4218-aaeb-bee781e9befb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.448839 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4zdg\" (UniqueName: \"kubernetes.io/projected/511d65e3-135b-4218-aaeb-bee781e9befb-kube-api-access-r4zdg\") pod \"nmstate-operator-5b5b58f5c8-xz2lq\" (UID: \"511d65e3-135b-4218-aaeb-bee781e9befb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.465206 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4zdg\" (UniqueName: \"kubernetes.io/projected/511d65e3-135b-4218-aaeb-bee781e9befb-kube-api-access-r4zdg\") pod \"nmstate-operator-5b5b58f5c8-xz2lq\" (UID: \"511d65e3-135b-4218-aaeb-bee781e9befb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.525072 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" Dec 02 13:15:32 crc kubenswrapper[4725]: I1202 13:15:32.952643 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq"] Dec 02 13:15:33 crc kubenswrapper[4725]: I1202 13:15:33.502421 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" event={"ID":"511d65e3-135b-4218-aaeb-bee781e9befb","Type":"ContainerStarted","Data":"2f1f767cebba756048bddbad62925ed94c501f349ce4e1c4d020ca8cda6dfa43"} Dec 02 13:15:35 crc kubenswrapper[4725]: I1202 13:15:35.512683 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" event={"ID":"511d65e3-135b-4218-aaeb-bee781e9befb","Type":"ContainerStarted","Data":"57c3f151c5ddf8da5682922b2fe2f326b4548ee214be806efa4fb12e4f9f5a90"} Dec 02 13:15:35 crc kubenswrapper[4725]: I1202 13:15:35.531860 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xz2lq" podStartSLOduration=1.125179456 podStartE2EDuration="3.531844079s" podCreationTimestamp="2025-12-02 13:15:32 +0000 UTC" firstStartedPulling="2025-12-02 13:15:32.959983282 +0000 UTC m=+663.916624977" lastFinishedPulling="2025-12-02 13:15:35.366647905 +0000 UTC m=+666.323289600" observedRunningTime="2025-12-02 13:15:35.529918891 +0000 UTC m=+666.486560586" watchObservedRunningTime="2025-12-02 13:15:35.531844079 +0000 UTC m=+666.488485774" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.217188 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.218270 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.220363 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-v8jbh" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.226407 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.227096 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.234530 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6cqs8"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.235399 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.235897 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.244324 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.287356 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.367040 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.367900 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.371797 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-s9vps" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.371983 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.372006 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.374943 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-ovs-socket\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.374982 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4fsq\" (UniqueName: \"kubernetes.io/projected/28a8d066-b52d-450e-a738-884b8b260da5-kube-api-access-d4fsq\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.375027 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-dbus-socket\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.375049 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-nmstate-lock\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.375065 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/15ebd594-c399-41a5-a80a-6ce365f9e657-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9rkbp\" (UID: \"15ebd594-c399-41a5-a80a-6ce365f9e657\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.375090 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwcqd\" (UniqueName: \"kubernetes.io/projected/15ebd594-c399-41a5-a80a-6ce365f9e657-kube-api-access-wwcqd\") pod \"nmstate-webhook-5f6d4c5ccb-9rkbp\" (UID: \"15ebd594-c399-41a5-a80a-6ce365f9e657\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.375111 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7l8z\" (UniqueName: \"kubernetes.io/projected/93447f0a-ea95-4ec5-beea-8733e261d5f3-kube-api-access-v7l8z\") pod \"nmstate-metrics-7f946cbc9-v7ngp\" (UID: \"93447f0a-ea95-4ec5-beea-8733e261d5f3\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.408147 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476394 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-ovs-socket\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476439 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4fsq\" (UniqueName: \"kubernetes.io/projected/28a8d066-b52d-450e-a738-884b8b260da5-kube-api-access-d4fsq\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476512 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-dbus-socket\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476532 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-ovs-socket\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476547 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-nmstate-lock\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476783 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4844270-686d-4f0b-aafc-5b2b1924444f-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.476968 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-nmstate-lock\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.477028 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/15ebd594-c399-41a5-a80a-6ce365f9e657-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9rkbp\" (UID: \"15ebd594-c399-41a5-a80a-6ce365f9e657\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.477100 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7phs\" (UniqueName: \"kubernetes.io/projected/a4844270-686d-4f0b-aafc-5b2b1924444f-kube-api-access-l7phs\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.477147 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a4844270-686d-4f0b-aafc-5b2b1924444f-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.477175 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwcqd\" (UniqueName: \"kubernetes.io/projected/15ebd594-c399-41a5-a80a-6ce365f9e657-kube-api-access-wwcqd\") pod \"nmstate-webhook-5f6d4c5ccb-9rkbp\" (UID: \"15ebd594-c399-41a5-a80a-6ce365f9e657\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.477181 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/28a8d066-b52d-450e-a738-884b8b260da5-dbus-socket\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.477229 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7l8z\" (UniqueName: \"kubernetes.io/projected/93447f0a-ea95-4ec5-beea-8733e261d5f3-kube-api-access-v7l8z\") pod \"nmstate-metrics-7f946cbc9-v7ngp\" (UID: \"93447f0a-ea95-4ec5-beea-8733e261d5f3\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.490116 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/15ebd594-c399-41a5-a80a-6ce365f9e657-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-9rkbp\" (UID: \"15ebd594-c399-41a5-a80a-6ce365f9e657\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.496394 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwcqd\" (UniqueName: \"kubernetes.io/projected/15ebd594-c399-41a5-a80a-6ce365f9e657-kube-api-access-wwcqd\") pod \"nmstate-webhook-5f6d4c5ccb-9rkbp\" (UID: \"15ebd594-c399-41a5-a80a-6ce365f9e657\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.497190 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4fsq\" (UniqueName: \"kubernetes.io/projected/28a8d066-b52d-450e-a738-884b8b260da5-kube-api-access-d4fsq\") pod \"nmstate-handler-6cqs8\" (UID: \"28a8d066-b52d-450e-a738-884b8b260da5\") " pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.497728 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7l8z\" (UniqueName: \"kubernetes.io/projected/93447f0a-ea95-4ec5-beea-8733e261d5f3-kube-api-access-v7l8z\") pod \"nmstate-metrics-7f946cbc9-v7ngp\" (UID: \"93447f0a-ea95-4ec5-beea-8733e261d5f3\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.535105 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.549837 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.571083 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-546b4d87bd-lk5ll"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.571323 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.571969 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.578314 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4844270-686d-4f0b-aafc-5b2b1924444f-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.578357 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7phs\" (UniqueName: \"kubernetes.io/projected/a4844270-686d-4f0b-aafc-5b2b1924444f-kube-api-access-l7phs\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.578383 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a4844270-686d-4f0b-aafc-5b2b1924444f-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: E1202 13:15:41.578514 4725 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 02 13:15:41 crc kubenswrapper[4725]: E1202 13:15:41.578682 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4844270-686d-4f0b-aafc-5b2b1924444f-plugin-serving-cert podName:a4844270-686d-4f0b-aafc-5b2b1924444f nodeName:}" failed. No retries permitted until 2025-12-02 13:15:42.078662059 +0000 UTC m=+673.035303794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/a4844270-686d-4f0b-aafc-5b2b1924444f-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-5dmqd" (UID: "a4844270-686d-4f0b-aafc-5b2b1924444f") : secret "plugin-serving-cert" not found Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.579347 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a4844270-686d-4f0b-aafc-5b2b1924444f-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.588003 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-546b4d87bd-lk5ll"] Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.618433 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7phs\" (UniqueName: \"kubernetes.io/projected/a4844270-686d-4f0b-aafc-5b2b1924444f-kube-api-access-l7phs\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679777 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-oauth-serving-cert\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679808 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-serving-cert\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679835 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sv6c\" (UniqueName: \"kubernetes.io/projected/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-kube-api-access-6sv6c\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679870 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-config\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679917 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-service-ca\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679944 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-oauth-config\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.679964 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-trusted-ca-bundle\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780671 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-service-ca\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780731 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-oauth-config\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780802 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-trusted-ca-bundle\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780827 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-oauth-serving-cert\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780854 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-serving-cert\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780882 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sv6c\" (UniqueName: \"kubernetes.io/projected/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-kube-api-access-6sv6c\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.780939 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-config\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.781787 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-service-ca\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.781907 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-config\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.782571 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-oauth-serving-cert\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.784640 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-trusted-ca-bundle\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.787096 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-oauth-config\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.787617 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-console-serving-cert\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.788829 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp"] Dec 02 13:15:41 crc kubenswrapper[4725]: W1202 13:15:41.794937 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15ebd594_c399_41a5_a80a_6ce365f9e657.slice/crio-4fc1b2319c9e72648db2522a991374bdb89c7fbaa9b30e3617a33da05700c531 WatchSource:0}: Error finding container 4fc1b2319c9e72648db2522a991374bdb89c7fbaa9b30e3617a33da05700c531: Status 404 returned error can't find the container with id 4fc1b2319c9e72648db2522a991374bdb89c7fbaa9b30e3617a33da05700c531 Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.802690 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sv6c\" (UniqueName: \"kubernetes.io/projected/b85b39c1-2b45-4cd1-afc3-1bb4c781806b-kube-api-access-6sv6c\") pod \"console-546b4d87bd-lk5ll\" (UID: \"b85b39c1-2b45-4cd1-afc3-1bb4c781806b\") " pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:41 crc kubenswrapper[4725]: I1202 13:15:41.940169 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.043775 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp"] Dec 02 13:15:42 crc kubenswrapper[4725]: W1202 13:15:42.047879 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93447f0a_ea95_4ec5_beea_8733e261d5f3.slice/crio-420b702f0af94db6d55f629fe597b99f9862a75c61dea3fbaf2f796966430ddc WatchSource:0}: Error finding container 420b702f0af94db6d55f629fe597b99f9862a75c61dea3fbaf2f796966430ddc: Status 404 returned error can't find the container with id 420b702f0af94db6d55f629fe597b99f9862a75c61dea3fbaf2f796966430ddc Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.084521 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4844270-686d-4f0b-aafc-5b2b1924444f-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.088446 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4844270-686d-4f0b-aafc-5b2b1924444f-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5dmqd\" (UID: \"a4844270-686d-4f0b-aafc-5b2b1924444f\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.109924 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-546b4d87bd-lk5ll"] Dec 02 13:15:42 crc kubenswrapper[4725]: W1202 13:15:42.113969 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb85b39c1_2b45_4cd1_afc3_1bb4c781806b.slice/crio-3edc414ea0acddb99e732bce0cbe9e6e9e941999ed47ee219adcec80c89adad5 WatchSource:0}: Error finding container 3edc414ea0acddb99e732bce0cbe9e6e9e941999ed47ee219adcec80c89adad5: Status 404 returned error can't find the container with id 3edc414ea0acddb99e732bce0cbe9e6e9e941999ed47ee219adcec80c89adad5 Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.302818 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.471615 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd"] Dec 02 13:15:42 crc kubenswrapper[4725]: W1202 13:15:42.478956 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4844270_686d_4f0b_aafc_5b2b1924444f.slice/crio-ba71c164475d39b261c46d6c75992a902d61a29628fd91ca33a4f102a9154a5e WatchSource:0}: Error finding container ba71c164475d39b261c46d6c75992a902d61a29628fd91ca33a4f102a9154a5e: Status 404 returned error can't find the container with id ba71c164475d39b261c46d6c75992a902d61a29628fd91ca33a4f102a9154a5e Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.558531 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" event={"ID":"15ebd594-c399-41a5-a80a-6ce365f9e657","Type":"ContainerStarted","Data":"4fc1b2319c9e72648db2522a991374bdb89c7fbaa9b30e3617a33da05700c531"} Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.559732 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6cqs8" event={"ID":"28a8d066-b52d-450e-a738-884b8b260da5","Type":"ContainerStarted","Data":"1f061ba0adc66b361a120aa048a82a589ad3e0c8bfbfe981d2fa9af5324e94b0"} Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.561563 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" event={"ID":"93447f0a-ea95-4ec5-beea-8733e261d5f3","Type":"ContainerStarted","Data":"420b702f0af94db6d55f629fe597b99f9862a75c61dea3fbaf2f796966430ddc"} Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.563541 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" event={"ID":"a4844270-686d-4f0b-aafc-5b2b1924444f","Type":"ContainerStarted","Data":"ba71c164475d39b261c46d6c75992a902d61a29628fd91ca33a4f102a9154a5e"} Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.565100 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-546b4d87bd-lk5ll" event={"ID":"b85b39c1-2b45-4cd1-afc3-1bb4c781806b","Type":"ContainerStarted","Data":"da3168a0b618ee2d2395d16121d8ead9d8313a0511d1689ace967c36d62b04e9"} Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.565159 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-546b4d87bd-lk5ll" event={"ID":"b85b39c1-2b45-4cd1-afc3-1bb4c781806b","Type":"ContainerStarted","Data":"3edc414ea0acddb99e732bce0cbe9e6e9e941999ed47ee219adcec80c89adad5"} Dec 02 13:15:42 crc kubenswrapper[4725]: I1202 13:15:42.585749 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-546b4d87bd-lk5ll" podStartSLOduration=1.58569776 podStartE2EDuration="1.58569776s" podCreationTimestamp="2025-12-02 13:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:15:42.579783233 +0000 UTC m=+673.536424928" watchObservedRunningTime="2025-12-02 13:15:42.58569776 +0000 UTC m=+673.542339465" Dec 02 13:15:44 crc kubenswrapper[4725]: I1202 13:15:44.578185 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" event={"ID":"15ebd594-c399-41a5-a80a-6ce365f9e657","Type":"ContainerStarted","Data":"ff40300a1ace60a6aac3bcc60d32b0b9981775ce47f5910ffbfbf7bf6b18a9b2"} Dec 02 13:15:44 crc kubenswrapper[4725]: I1202 13:15:44.578903 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:15:44 crc kubenswrapper[4725]: I1202 13:15:44.580235 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6cqs8" event={"ID":"28a8d066-b52d-450e-a738-884b8b260da5","Type":"ContainerStarted","Data":"bb7d8bc6c0cc971fb3c207fd49ffcfa2bd05f45ff9aa85431ba42c7b7f8113c5"} Dec 02 13:15:44 crc kubenswrapper[4725]: I1202 13:15:44.580288 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:44 crc kubenswrapper[4725]: I1202 13:15:44.582043 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" event={"ID":"93447f0a-ea95-4ec5-beea-8733e261d5f3","Type":"ContainerStarted","Data":"7c4ab3a526b320dfd859ff939b3430669195a6f1058cbdf49c41635b21fef488"} Dec 02 13:15:44 crc kubenswrapper[4725]: I1202 13:15:44.631594 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" podStartSLOduration=1.25323566 podStartE2EDuration="3.631563827s" podCreationTimestamp="2025-12-02 13:15:41 +0000 UTC" firstStartedPulling="2025-12-02 13:15:41.799484169 +0000 UTC m=+672.756125864" lastFinishedPulling="2025-12-02 13:15:44.177812336 +0000 UTC m=+675.134454031" observedRunningTime="2025-12-02 13:15:44.607232721 +0000 UTC m=+675.563874426" watchObservedRunningTime="2025-12-02 13:15:44.631563827 +0000 UTC m=+675.588205522" Dec 02 13:15:45 crc kubenswrapper[4725]: I1202 13:15:45.590080 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" event={"ID":"a4844270-686d-4f0b-aafc-5b2b1924444f","Type":"ContainerStarted","Data":"12f32c1da1e3f60b88199317893e3f888e814e784cea543320b50128210f07b0"} Dec 02 13:15:45 crc kubenswrapper[4725]: I1202 13:15:45.608125 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6cqs8" podStartSLOduration=2.066413104 podStartE2EDuration="4.608101109s" podCreationTimestamp="2025-12-02 13:15:41 +0000 UTC" firstStartedPulling="2025-12-02 13:15:41.635029063 +0000 UTC m=+672.591670758" lastFinishedPulling="2025-12-02 13:15:44.176717068 +0000 UTC m=+675.133358763" observedRunningTime="2025-12-02 13:15:44.632025529 +0000 UTC m=+675.588667224" watchObservedRunningTime="2025-12-02 13:15:45.608101109 +0000 UTC m=+676.564742804" Dec 02 13:15:45 crc kubenswrapper[4725]: I1202 13:15:45.611561 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5dmqd" podStartSLOduration=1.8305736179999998 podStartE2EDuration="4.611552895s" podCreationTimestamp="2025-12-02 13:15:41 +0000 UTC" firstStartedPulling="2025-12-02 13:15:42.481759761 +0000 UTC m=+673.438401456" lastFinishedPulling="2025-12-02 13:15:45.262739038 +0000 UTC m=+676.219380733" observedRunningTime="2025-12-02 13:15:45.604227843 +0000 UTC m=+676.560869558" watchObservedRunningTime="2025-12-02 13:15:45.611552895 +0000 UTC m=+676.568194580" Dec 02 13:15:47 crc kubenswrapper[4725]: I1202 13:15:47.604019 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" event={"ID":"93447f0a-ea95-4ec5-beea-8733e261d5f3","Type":"ContainerStarted","Data":"3a503464edabed03197a9bd749df887075bdc6d84ea9432a5b8b001b870ce326"} Dec 02 13:15:47 crc kubenswrapper[4725]: I1202 13:15:47.631063 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-v7ngp" podStartSLOduration=2.176608678 podStartE2EDuration="6.631047166s" podCreationTimestamp="2025-12-02 13:15:41 +0000 UTC" firstStartedPulling="2025-12-02 13:15:42.051391742 +0000 UTC m=+673.008033437" lastFinishedPulling="2025-12-02 13:15:46.50583022 +0000 UTC m=+677.462471925" observedRunningTime="2025-12-02 13:15:47.622275157 +0000 UTC m=+678.578916892" watchObservedRunningTime="2025-12-02 13:15:47.631047166 +0000 UTC m=+678.587688861" Dec 02 13:15:51 crc kubenswrapper[4725]: I1202 13:15:51.591176 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6cqs8" Dec 02 13:15:51 crc kubenswrapper[4725]: I1202 13:15:51.940795 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:51 crc kubenswrapper[4725]: I1202 13:15:51.940981 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:51 crc kubenswrapper[4725]: I1202 13:15:51.948065 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:52 crc kubenswrapper[4725]: I1202 13:15:52.640036 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-546b4d87bd-lk5ll" Dec 02 13:15:52 crc kubenswrapper[4725]: I1202 13:15:52.730277 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7z7tx"] Dec 02 13:15:54 crc kubenswrapper[4725]: I1202 13:15:54.284828 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:15:54 crc kubenswrapper[4725]: I1202 13:15:54.285191 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:16:01 crc kubenswrapper[4725]: I1202 13:16:01.558527 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-9rkbp" Dec 02 13:16:12 crc kubenswrapper[4725]: I1202 13:16:12.960173 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr"] Dec 02 13:16:12 crc kubenswrapper[4725]: I1202 13:16:12.962728 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:12 crc kubenswrapper[4725]: I1202 13:16:12.964743 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 13:16:12 crc kubenswrapper[4725]: I1202 13:16:12.972593 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr"] Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.003373 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gk7j\" (UniqueName: \"kubernetes.io/projected/a9565732-da1b-4ccc-8913-fd706474ce9a-kube-api-access-7gk7j\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.003781 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.003832 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.104684 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.104751 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gk7j\" (UniqueName: \"kubernetes.io/projected/a9565732-da1b-4ccc-8913-fd706474ce9a-kube-api-access-7gk7j\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.104801 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.105177 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.105186 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.132283 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gk7j\" (UniqueName: \"kubernetes.io/projected/a9565732-da1b-4ccc-8913-fd706474ce9a-kube-api-access-7gk7j\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.278302 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.666424 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr"] Dec 02 13:16:13 crc kubenswrapper[4725]: I1202 13:16:13.755140 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" event={"ID":"a9565732-da1b-4ccc-8913-fd706474ce9a","Type":"ContainerStarted","Data":"6221c8bb0ed3f19eab5cfa7924a93483c1d8ed4185c5823d3389a3ce57918409"} Dec 02 13:16:14 crc kubenswrapper[4725]: I1202 13:16:14.763842 4725 generic.go:334] "Generic (PLEG): container finished" podID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerID="8f13bc2f4a1358af1fc332b2ea86f0918225acfd047ee02bdb14581e87750c59" exitCode=0 Dec 02 13:16:14 crc kubenswrapper[4725]: I1202 13:16:14.763928 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" event={"ID":"a9565732-da1b-4ccc-8913-fd706474ce9a","Type":"ContainerDied","Data":"8f13bc2f4a1358af1fc332b2ea86f0918225acfd047ee02bdb14581e87750c59"} Dec 02 13:16:16 crc kubenswrapper[4725]: I1202 13:16:16.783312 4725 generic.go:334] "Generic (PLEG): container finished" podID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerID="cbe1cfa6c37874742b1683d6fd7e8e665f5fb5a01f62b10a1ea187a819b7a523" exitCode=0 Dec 02 13:16:16 crc kubenswrapper[4725]: I1202 13:16:16.783364 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" event={"ID":"a9565732-da1b-4ccc-8913-fd706474ce9a","Type":"ContainerDied","Data":"cbe1cfa6c37874742b1683d6fd7e8e665f5fb5a01f62b10a1ea187a819b7a523"} Dec 02 13:16:17 crc kubenswrapper[4725]: I1202 13:16:17.782334 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7z7tx" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" containerName="console" containerID="cri-o://0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58" gracePeriod=15 Dec 02 13:16:17 crc kubenswrapper[4725]: I1202 13:16:17.792813 4725 generic.go:334] "Generic (PLEG): container finished" podID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerID="60be02b229e6a15bfbca17a28d919290b7d7543fbacdaa58b43b9070e53ec200" exitCode=0 Dec 02 13:16:17 crc kubenswrapper[4725]: I1202 13:16:17.792854 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" event={"ID":"a9565732-da1b-4ccc-8913-fd706474ce9a","Type":"ContainerDied","Data":"60be02b229e6a15bfbca17a28d919290b7d7543fbacdaa58b43b9070e53ec200"} Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.085648 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7z7tx_7971baa9-a25e-47e2-ac21-8c89db535872/console/0.log" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.085957 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164431 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-serving-cert\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164523 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-trusted-ca-bundle\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164550 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lzdh\" (UniqueName: \"kubernetes.io/projected/7971baa9-a25e-47e2-ac21-8c89db535872-kube-api-access-2lzdh\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164567 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-console-config\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164607 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-service-ca\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164645 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-oauth-config\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.164671 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-oauth-serving-cert\") pod \"7971baa9-a25e-47e2-ac21-8c89db535872\" (UID: \"7971baa9-a25e-47e2-ac21-8c89db535872\") " Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.165218 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-console-config" (OuterVolumeSpecName: "console-config") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.165226 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-service-ca" (OuterVolumeSpecName: "service-ca") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.165641 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.165738 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.169449 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.169588 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.170345 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7971baa9-a25e-47e2-ac21-8c89db535872-kube-api-access-2lzdh" (OuterVolumeSpecName: "kube-api-access-2lzdh") pod "7971baa9-a25e-47e2-ac21-8c89db535872" (UID: "7971baa9-a25e-47e2-ac21-8c89db535872"). InnerVolumeSpecName "kube-api-access-2lzdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266144 4725 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266196 4725 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266209 4725 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266218 4725 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7971baa9-a25e-47e2-ac21-8c89db535872-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266227 4725 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266236 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lzdh\" (UniqueName: \"kubernetes.io/projected/7971baa9-a25e-47e2-ac21-8c89db535872-kube-api-access-2lzdh\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.266244 4725 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7971baa9-a25e-47e2-ac21-8c89db535872-console-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.801699 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7z7tx_7971baa9-a25e-47e2-ac21-8c89db535872/console/0.log" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.801768 4725 generic.go:334] "Generic (PLEG): container finished" podID="7971baa9-a25e-47e2-ac21-8c89db535872" containerID="0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58" exitCode=2 Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.801926 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7z7tx" event={"ID":"7971baa9-a25e-47e2-ac21-8c89db535872","Type":"ContainerDied","Data":"0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58"} Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.801988 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7z7tx" event={"ID":"7971baa9-a25e-47e2-ac21-8c89db535872","Type":"ContainerDied","Data":"3bab371b695af668c432a4add5d8efe243d633031b9deacdd74b81ed99e0352e"} Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.802001 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7z7tx" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.802008 4725 scope.go:117] "RemoveContainer" containerID="0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.818824 4725 scope.go:117] "RemoveContainer" containerID="0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58" Dec 02 13:16:18 crc kubenswrapper[4725]: E1202 13:16:18.819200 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58\": container with ID starting with 0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58 not found: ID does not exist" containerID="0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.819261 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58"} err="failed to get container status \"0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58\": rpc error: code = NotFound desc = could not find container \"0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58\": container with ID starting with 0b8406e185b6a247c880633b18d487c16f59eb896a802741288b27a0ed385f58 not found: ID does not exist" Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.844653 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7z7tx"] Dec 02 13:16:18 crc kubenswrapper[4725]: I1202 13:16:18.848420 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7z7tx"] Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.071298 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.075013 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gk7j\" (UniqueName: \"kubernetes.io/projected/a9565732-da1b-4ccc-8913-fd706474ce9a-kube-api-access-7gk7j\") pod \"a9565732-da1b-4ccc-8913-fd706474ce9a\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.079999 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9565732-da1b-4ccc-8913-fd706474ce9a-kube-api-access-7gk7j" (OuterVolumeSpecName: "kube-api-access-7gk7j") pod "a9565732-da1b-4ccc-8913-fd706474ce9a" (UID: "a9565732-da1b-4ccc-8913-fd706474ce9a"). InnerVolumeSpecName "kube-api-access-7gk7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.175828 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-bundle\") pod \"a9565732-da1b-4ccc-8913-fd706474ce9a\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.175892 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-util\") pod \"a9565732-da1b-4ccc-8913-fd706474ce9a\" (UID: \"a9565732-da1b-4ccc-8913-fd706474ce9a\") " Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.176103 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gk7j\" (UniqueName: \"kubernetes.io/projected/a9565732-da1b-4ccc-8913-fd706474ce9a-kube-api-access-7gk7j\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.176822 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-bundle" (OuterVolumeSpecName: "bundle") pod "a9565732-da1b-4ccc-8913-fd706474ce9a" (UID: "a9565732-da1b-4ccc-8913-fd706474ce9a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.274286 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" path="/var/lib/kubelet/pods/7971baa9-a25e-47e2-ac21-8c89db535872/volumes" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.277020 4725 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.293095 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-util" (OuterVolumeSpecName: "util") pod "a9565732-da1b-4ccc-8913-fd706474ce9a" (UID: "a9565732-da1b-4ccc-8913-fd706474ce9a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.378387 4725 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a9565732-da1b-4ccc-8913-fd706474ce9a-util\") on node \"crc\" DevicePath \"\"" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.809568 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" event={"ID":"a9565732-da1b-4ccc-8913-fd706474ce9a","Type":"ContainerDied","Data":"6221c8bb0ed3f19eab5cfa7924a93483c1d8ed4185c5823d3389a3ce57918409"} Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.809602 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6221c8bb0ed3f19eab5cfa7924a93483c1d8ed4185c5823d3389a3ce57918409" Dec 02 13:16:19 crc kubenswrapper[4725]: I1202 13:16:19.809656 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr" Dec 02 13:16:24 crc kubenswrapper[4725]: I1202 13:16:24.284578 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:16:24 crc kubenswrapper[4725]: I1202 13:16:24.284936 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.695825 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-747844468f-qq5ns"] Dec 02 13:16:28 crc kubenswrapper[4725]: E1202 13:16:28.696361 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="extract" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.696376 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="extract" Dec 02 13:16:28 crc kubenswrapper[4725]: E1202 13:16:28.696387 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="pull" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.696395 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="pull" Dec 02 13:16:28 crc kubenswrapper[4725]: E1202 13:16:28.696409 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="util" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.696417 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="util" Dec 02 13:16:28 crc kubenswrapper[4725]: E1202 13:16:28.696437 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" containerName="console" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.696444 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" containerName="console" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.696569 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7971baa9-a25e-47e2-ac21-8c89db535872" containerName="console" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.696582 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9565732-da1b-4ccc-8913-fd706474ce9a" containerName="extract" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.697050 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.699116 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.699649 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.699749 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.699798 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.699915 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-mcqdl" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.714860 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-747844468f-qq5ns"] Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.885855 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7l8f\" (UniqueName: \"kubernetes.io/projected/bc5b46e6-ecea-4260-a386-e7408c29a842-kube-api-access-t7l8f\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.885952 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc5b46e6-ecea-4260-a386-e7408c29a842-apiservice-cert\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.885974 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc5b46e6-ecea-4260-a386-e7408c29a842-webhook-cert\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.918326 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q"] Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.918970 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.921796 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.921951 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qkp99" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.921964 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.939900 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q"] Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.986766 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-apiservice-cert\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.986822 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lhxq\" (UniqueName: \"kubernetes.io/projected/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-kube-api-access-2lhxq\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.986952 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc5b46e6-ecea-4260-a386-e7408c29a842-apiservice-cert\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.986993 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc5b46e6-ecea-4260-a386-e7408c29a842-webhook-cert\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.987031 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-webhook-cert\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.987065 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7l8f\" (UniqueName: \"kubernetes.io/projected/bc5b46e6-ecea-4260-a386-e7408c29a842-kube-api-access-t7l8f\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:28 crc kubenswrapper[4725]: I1202 13:16:28.992392 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc5b46e6-ecea-4260-a386-e7408c29a842-apiservice-cert\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.008225 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc5b46e6-ecea-4260-a386-e7408c29a842-webhook-cert\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.013086 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7l8f\" (UniqueName: \"kubernetes.io/projected/bc5b46e6-ecea-4260-a386-e7408c29a842-kube-api-access-t7l8f\") pod \"metallb-operator-controller-manager-747844468f-qq5ns\" (UID: \"bc5b46e6-ecea-4260-a386-e7408c29a842\") " pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.087407 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-webhook-cert\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.087477 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-apiservice-cert\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.087510 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lhxq\" (UniqueName: \"kubernetes.io/projected/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-kube-api-access-2lhxq\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.093103 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-webhook-cert\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.098882 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-apiservice-cert\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.105147 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lhxq\" (UniqueName: \"kubernetes.io/projected/2315c4fc-8de9-4d85-9fab-a2cedd0eb95d-kube-api-access-2lhxq\") pod \"metallb-operator-webhook-server-857fd986b8-tqd7q\" (UID: \"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d\") " pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.237421 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-qkp99" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.243609 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.318872 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-mcqdl" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.326384 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.758063 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q"] Dec 02 13:16:29 crc kubenswrapper[4725]: W1202 13:16:29.763835 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2315c4fc_8de9_4d85_9fab_a2cedd0eb95d.slice/crio-80db2faedcb7cfe89fe9910a606b60855ec451798c75c2182e7e0c5a059c98be WatchSource:0}: Error finding container 80db2faedcb7cfe89fe9910a606b60855ec451798c75c2182e7e0c5a059c98be: Status 404 returned error can't find the container with id 80db2faedcb7cfe89fe9910a606b60855ec451798c75c2182e7e0c5a059c98be Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.857607 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-747844468f-qq5ns"] Dec 02 13:16:29 crc kubenswrapper[4725]: I1202 13:16:29.858179 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" event={"ID":"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d","Type":"ContainerStarted","Data":"80db2faedcb7cfe89fe9910a606b60855ec451798c75c2182e7e0c5a059c98be"} Dec 02 13:16:29 crc kubenswrapper[4725]: W1202 13:16:29.861495 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc5b46e6_ecea_4260_a386_e7408c29a842.slice/crio-359dc29fa040a0b2fa235b3dc6261aab9a18da4cbaef7879838880cff02de6da WatchSource:0}: Error finding container 359dc29fa040a0b2fa235b3dc6261aab9a18da4cbaef7879838880cff02de6da: Status 404 returned error can't find the container with id 359dc29fa040a0b2fa235b3dc6261aab9a18da4cbaef7879838880cff02de6da Dec 02 13:16:30 crc kubenswrapper[4725]: I1202 13:16:30.864758 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" event={"ID":"bc5b46e6-ecea-4260-a386-e7408c29a842","Type":"ContainerStarted","Data":"359dc29fa040a0b2fa235b3dc6261aab9a18da4cbaef7879838880cff02de6da"} Dec 02 13:16:36 crc kubenswrapper[4725]: I1202 13:16:36.907644 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" event={"ID":"2315c4fc-8de9-4d85-9fab-a2cedd0eb95d","Type":"ContainerStarted","Data":"055b30a963519e770a8d206b050781a629e913636ce93154300e1f7236b3c04e"} Dec 02 13:16:36 crc kubenswrapper[4725]: I1202 13:16:36.908227 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:36 crc kubenswrapper[4725]: I1202 13:16:36.909668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" event={"ID":"bc5b46e6-ecea-4260-a386-e7408c29a842","Type":"ContainerStarted","Data":"f53d94d421762d9eceef7c2534da92efe2fd9e45ff7eff93e355529ee6677265"} Dec 02 13:16:36 crc kubenswrapper[4725]: I1202 13:16:36.909835 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:16:36 crc kubenswrapper[4725]: I1202 13:16:36.928720 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" podStartSLOduration=2.797473385 podStartE2EDuration="8.928698016s" podCreationTimestamp="2025-12-02 13:16:28 +0000 UTC" firstStartedPulling="2025-12-02 13:16:29.770020584 +0000 UTC m=+720.726662269" lastFinishedPulling="2025-12-02 13:16:35.901245215 +0000 UTC m=+726.857886900" observedRunningTime="2025-12-02 13:16:36.924281686 +0000 UTC m=+727.880923411" watchObservedRunningTime="2025-12-02 13:16:36.928698016 +0000 UTC m=+727.885339711" Dec 02 13:16:36 crc kubenswrapper[4725]: I1202 13:16:36.949547 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" podStartSLOduration=2.940274791 podStartE2EDuration="8.949524084s" podCreationTimestamp="2025-12-02 13:16:28 +0000 UTC" firstStartedPulling="2025-12-02 13:16:29.864052376 +0000 UTC m=+720.820694061" lastFinishedPulling="2025-12-02 13:16:35.873301659 +0000 UTC m=+726.829943354" observedRunningTime="2025-12-02 13:16:36.946636593 +0000 UTC m=+727.903278308" watchObservedRunningTime="2025-12-02 13:16:36.949524084 +0000 UTC m=+727.906165769" Dec 02 13:16:49 crc kubenswrapper[4725]: I1202 13:16:49.249742 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-857fd986b8-tqd7q" Dec 02 13:16:54 crc kubenswrapper[4725]: I1202 13:16:54.284491 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:16:54 crc kubenswrapper[4725]: I1202 13:16:54.284955 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:16:54 crc kubenswrapper[4725]: I1202 13:16:54.284993 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:16:54 crc kubenswrapper[4725]: I1202 13:16:54.285491 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"01e784ee70283b45e319c7d0015626dcb847bd5037e05305c113f2ede968d762"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:16:54 crc kubenswrapper[4725]: I1202 13:16:54.285535 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://01e784ee70283b45e319c7d0015626dcb847bd5037e05305c113f2ede968d762" gracePeriod=600 Dec 02 13:16:56 crc kubenswrapper[4725]: I1202 13:16:56.005121 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="01e784ee70283b45e319c7d0015626dcb847bd5037e05305c113f2ede968d762" exitCode=0 Dec 02 13:16:56 crc kubenswrapper[4725]: I1202 13:16:56.005206 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"01e784ee70283b45e319c7d0015626dcb847bd5037e05305c113f2ede968d762"} Dec 02 13:16:56 crc kubenswrapper[4725]: I1202 13:16:56.005423 4725 scope.go:117] "RemoveContainer" containerID="7a038ec0fa554f0a9752863aee67be397498e9cc4ec37228c8b10ecc3fbb0599" Dec 02 13:16:57 crc kubenswrapper[4725]: I1202 13:16:57.012122 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"203c78b7d89a690dfa6df4df74f2999e81c6261249dd7853e1b1c46205cb6197"} Dec 02 13:17:01 crc kubenswrapper[4725]: I1202 13:17:01.436565 4725 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.198668 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfj5"] Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.200648 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.210620 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfj5"] Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.328951 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-catalog-content\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.329050 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-utilities\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.329192 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v272\" (UniqueName: \"kubernetes.io/projected/ccfb8381-ea5f-4336-b453-0cb863e6ec96-kube-api-access-6v272\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.430022 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-catalog-content\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.430090 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-utilities\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.430267 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v272\" (UniqueName: \"kubernetes.io/projected/ccfb8381-ea5f-4336-b453-0cb863e6ec96-kube-api-access-6v272\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.431419 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-utilities\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.431834 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-catalog-content\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.456011 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v272\" (UniqueName: \"kubernetes.io/projected/ccfb8381-ea5f-4336-b453-0cb863e6ec96-kube-api-access-6v272\") pod \"redhat-marketplace-kgfj5\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.553564 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:05 crc kubenswrapper[4725]: I1202 13:17:05.771823 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfj5"] Dec 02 13:17:05 crc kubenswrapper[4725]: W1202 13:17:05.772016 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccfb8381_ea5f_4336_b453_0cb863e6ec96.slice/crio-67dd0c081812dead9c800fabf7a0fbbc9692d36e38c88e00d21c683818031905 WatchSource:0}: Error finding container 67dd0c081812dead9c800fabf7a0fbbc9692d36e38c88e00d21c683818031905: Status 404 returned error can't find the container with id 67dd0c081812dead9c800fabf7a0fbbc9692d36e38c88e00d21c683818031905 Dec 02 13:17:06 crc kubenswrapper[4725]: I1202 13:17:06.071345 4725 generic.go:334] "Generic (PLEG): container finished" podID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerID="cf4984f621117916cc803b68337773004015a6933f5bb2a15a1c65cab46be75a" exitCode=0 Dec 02 13:17:06 crc kubenswrapper[4725]: I1202 13:17:06.071396 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfj5" event={"ID":"ccfb8381-ea5f-4336-b453-0cb863e6ec96","Type":"ContainerDied","Data":"cf4984f621117916cc803b68337773004015a6933f5bb2a15a1c65cab46be75a"} Dec 02 13:17:06 crc kubenswrapper[4725]: I1202 13:17:06.071441 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfj5" event={"ID":"ccfb8381-ea5f-4336-b453-0cb863e6ec96","Type":"ContainerStarted","Data":"67dd0c081812dead9c800fabf7a0fbbc9692d36e38c88e00d21c683818031905"} Dec 02 13:17:08 crc kubenswrapper[4725]: I1202 13:17:08.083294 4725 generic.go:334] "Generic (PLEG): container finished" podID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerID="884a257ed1be387eba17185dd00a07b5ed1cf23cf38e495fd52d378b87b1b7f6" exitCode=0 Dec 02 13:17:08 crc kubenswrapper[4725]: I1202 13:17:08.083348 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfj5" event={"ID":"ccfb8381-ea5f-4336-b453-0cb863e6ec96","Type":"ContainerDied","Data":"884a257ed1be387eba17185dd00a07b5ed1cf23cf38e495fd52d378b87b1b7f6"} Dec 02 13:17:09 crc kubenswrapper[4725]: I1202 13:17:09.098836 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfj5" event={"ID":"ccfb8381-ea5f-4336-b453-0cb863e6ec96","Type":"ContainerStarted","Data":"12c1ef0c2cbd247d6f1c75cb6124975aed8a1b4459c1274e5203177df08a929d"} Dec 02 13:17:09 crc kubenswrapper[4725]: I1202 13:17:09.122213 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kgfj5" podStartSLOduration=1.670112838 podStartE2EDuration="4.122188532s" podCreationTimestamp="2025-12-02 13:17:05 +0000 UTC" firstStartedPulling="2025-12-02 13:17:06.072968255 +0000 UTC m=+757.029609950" lastFinishedPulling="2025-12-02 13:17:08.525043949 +0000 UTC m=+759.481685644" observedRunningTime="2025-12-02 13:17:09.116197683 +0000 UTC m=+760.072839418" watchObservedRunningTime="2025-12-02 13:17:09.122188532 +0000 UTC m=+760.078830257" Dec 02 13:17:09 crc kubenswrapper[4725]: I1202 13:17:09.329106 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-747844468f-qq5ns" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.085151 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-bspx6"] Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.087804 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.091614 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.091761 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46"] Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.092748 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.094027 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/46316422-5e47-461c-8827-6e12d896134c-frr-startup\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.094177 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46316422-5e47-461c-8827-6e12d896134c-metrics-certs\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.094330 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-frr-sockets\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.095135 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-frr-conf\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.095322 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc6t7\" (UniqueName: \"kubernetes.io/projected/46316422-5e47-461c-8827-6e12d896134c-kube-api-access-wc6t7\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.095499 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjgcs\" (UniqueName: \"kubernetes.io/projected/00168d15-ce5e-4164-891c-1403031bee1d-kube-api-access-sjgcs\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.095628 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-reloader\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.095760 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00168d15-ce5e-4164-891c-1403031bee1d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.095906 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-metrics\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.100581 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.102374 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.102378 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-x8zd8" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.106646 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46"] Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.184241 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-dhmrd"] Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.185215 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.186929 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.188792 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-72q7c" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.189208 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.189496 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197063 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197121 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc6t7\" (UniqueName: \"kubernetes.io/projected/46316422-5e47-461c-8827-6e12d896134c-kube-api-access-wc6t7\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197156 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjgcs\" (UniqueName: \"kubernetes.io/projected/00168d15-ce5e-4164-891c-1403031bee1d-kube-api-access-sjgcs\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197221 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-reloader\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197253 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00168d15-ce5e-4164-891c-1403031bee1d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197292 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-metrics\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197321 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/46316422-5e47-461c-8827-6e12d896134c-frr-startup\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197341 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46316422-5e47-461c-8827-6e12d896134c-metrics-certs\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197368 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-frr-sockets\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197404 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metrics-certs\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197430 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnbl9\" (UniqueName: \"kubernetes.io/projected/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-kube-api-access-vnbl9\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197500 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metallb-excludel2\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197531 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-frr-conf\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.197757 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-reloader\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.198009 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-metrics\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.198031 4725 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.198093 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-frr-sockets\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.198103 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/00168d15-ce5e-4164-891c-1403031bee1d-cert podName:00168d15-ce5e-4164-891c-1403031bee1d nodeName:}" failed. No retries permitted until 2025-12-02 13:17:10.69808214 +0000 UTC m=+761.654723935 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/00168d15-ce5e-4164-891c-1403031bee1d-cert") pod "frr-k8s-webhook-server-7fcb986d4-c7g46" (UID: "00168d15-ce5e-4164-891c-1403031bee1d") : secret "frr-k8s-webhook-server-cert" not found Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.198165 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/46316422-5e47-461c-8827-6e12d896134c-frr-conf\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.198270 4725 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.198414 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46316422-5e47-461c-8827-6e12d896134c-metrics-certs podName:46316422-5e47-461c-8827-6e12d896134c nodeName:}" failed. No retries permitted until 2025-12-02 13:17:10.698391097 +0000 UTC m=+761.655032892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/46316422-5e47-461c-8827-6e12d896134c-metrics-certs") pod "frr-k8s-bspx6" (UID: "46316422-5e47-461c-8827-6e12d896134c") : secret "frr-k8s-certs-secret" not found Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.198836 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/46316422-5e47-461c-8827-6e12d896134c-frr-startup\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.202674 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-bkrzn"] Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.203537 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.208396 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.218947 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-bkrzn"] Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.263797 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjgcs\" (UniqueName: \"kubernetes.io/projected/00168d15-ce5e-4164-891c-1403031bee1d-kube-api-access-sjgcs\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.264684 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc6t7\" (UniqueName: \"kubernetes.io/projected/46316422-5e47-461c-8827-6e12d896134c-kube-api-access-wc6t7\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299108 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metrics-certs\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299164 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnbl9\" (UniqueName: \"kubernetes.io/projected/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-kube-api-access-vnbl9\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299203 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metallb-excludel2\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299235 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299285 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-cert\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299304 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk5qq\" (UniqueName: \"kubernetes.io/projected/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-kube-api-access-wk5qq\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.299316 4725 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.299419 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metrics-certs podName:ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb nodeName:}" failed. No retries permitted until 2025-12-02 13:17:10.799392243 +0000 UTC m=+761.756033938 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metrics-certs") pod "speaker-dhmrd" (UID: "ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb") : secret "speaker-certs-secret" not found Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.299337 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-metrics-certs\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.299619 4725 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.299661 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist podName:ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb nodeName:}" failed. No retries permitted until 2025-12-02 13:17:10.799653539 +0000 UTC m=+761.756295234 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist") pod "speaker-dhmrd" (UID: "ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb") : secret "metallb-memberlist" not found Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.300853 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metallb-excludel2\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.323254 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnbl9\" (UniqueName: \"kubernetes.io/projected/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-kube-api-access-vnbl9\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.400809 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-cert\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.400856 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk5qq\" (UniqueName: \"kubernetes.io/projected/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-kube-api-access-wk5qq\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.400884 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-metrics-certs\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.404044 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-metrics-certs\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.404887 4725 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.415338 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-cert\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.419109 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk5qq\" (UniqueName: \"kubernetes.io/projected/ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b-kube-api-access-wk5qq\") pod \"controller-f8648f98b-bkrzn\" (UID: \"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b\") " pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.516385 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.704765 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00168d15-ce5e-4164-891c-1403031bee1d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.704819 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46316422-5e47-461c-8827-6e12d896134c-metrics-certs\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.712141 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/46316422-5e47-461c-8827-6e12d896134c-metrics-certs\") pod \"frr-k8s-bspx6\" (UID: \"46316422-5e47-461c-8827-6e12d896134c\") " pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.712228 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/00168d15-ce5e-4164-891c-1403031bee1d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-c7g46\" (UID: \"00168d15-ce5e-4164-891c-1403031bee1d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.805390 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metrics-certs\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.805475 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.805597 4725 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 02 13:17:10 crc kubenswrapper[4725]: E1202 13:17:10.805652 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist podName:ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb nodeName:}" failed. No retries permitted until 2025-12-02 13:17:11.805638182 +0000 UTC m=+762.762279877 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist") pod "speaker-dhmrd" (UID: "ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb") : secret "metallb-memberlist" not found Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.809651 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-metrics-certs\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:10 crc kubenswrapper[4725]: I1202 13:17:10.916041 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-bkrzn"] Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.003865 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.011610 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.132678 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bkrzn" event={"ID":"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b","Type":"ContainerStarted","Data":"085560d8aa2d43e0492885b3c9617631b8c978656fcc77c0917a6a206342f54d"} Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.133029 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bkrzn" event={"ID":"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b","Type":"ContainerStarted","Data":"ba0a819082c5b20b43ead44710d58cd0ac94c88ee738667c0a11e42dd97c74e8"} Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.231088 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46"] Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.847543 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.855368 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb-memberlist\") pod \"speaker-dhmrd\" (UID: \"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb\") " pod="metallb-system/speaker-dhmrd" Dec 02 13:17:11 crc kubenswrapper[4725]: I1202 13:17:11.999189 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dhmrd" Dec 02 13:17:12 crc kubenswrapper[4725]: I1202 13:17:12.143387 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" event={"ID":"00168d15-ce5e-4164-891c-1403031bee1d","Type":"ContainerStarted","Data":"eb82f2e241bfd078005adcf000ca8e393869a1e0c3d38d645058cac21417d8be"} Dec 02 13:17:12 crc kubenswrapper[4725]: I1202 13:17:12.145185 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-bkrzn" event={"ID":"ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b","Type":"ContainerStarted","Data":"ae58e4917d5b6b7e3603da85e47fbaa6a45263688fce55d054c5959fd1a0fac3"} Dec 02 13:17:12 crc kubenswrapper[4725]: I1202 13:17:12.145520 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:12 crc kubenswrapper[4725]: I1202 13:17:12.146176 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dhmrd" event={"ID":"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb","Type":"ContainerStarted","Data":"470cc3f847dc8ae30db1343cb59300c28f355ae699e83ffec36fc5b0f71ac244"} Dec 02 13:17:12 crc kubenswrapper[4725]: I1202 13:17:12.147442 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"62f5ccf87e9906fe9837d4a4f5e3aae9ec41f2e79f49aa78b5a3bb932b568088"} Dec 02 13:17:12 crc kubenswrapper[4725]: I1202 13:17:12.159814 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-bkrzn" podStartSLOduration=2.15979525 podStartE2EDuration="2.15979525s" podCreationTimestamp="2025-12-02 13:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:17:12.158869886 +0000 UTC m=+763.115511581" watchObservedRunningTime="2025-12-02 13:17:12.15979525 +0000 UTC m=+763.116436935" Dec 02 13:17:13 crc kubenswrapper[4725]: I1202 13:17:13.162166 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dhmrd" event={"ID":"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb","Type":"ContainerStarted","Data":"a4614bd0435df9e39734909fcf4dc1d92cd15eebb43847b49350a834e8a584b5"} Dec 02 13:17:13 crc kubenswrapper[4725]: I1202 13:17:13.162515 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-dhmrd" Dec 02 13:17:13 crc kubenswrapper[4725]: I1202 13:17:13.162530 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dhmrd" event={"ID":"ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb","Type":"ContainerStarted","Data":"36b24d88f29a30610610620b83915ff0d043c0ed516586b8b4d2996d674d1207"} Dec 02 13:17:13 crc kubenswrapper[4725]: I1202 13:17:13.184730 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-dhmrd" podStartSLOduration=3.184710447 podStartE2EDuration="3.184710447s" podCreationTimestamp="2025-12-02 13:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:17:13.183152909 +0000 UTC m=+764.139794604" watchObservedRunningTime="2025-12-02 13:17:13.184710447 +0000 UTC m=+764.141352142" Dec 02 13:17:15 crc kubenswrapper[4725]: I1202 13:17:15.553918 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:15 crc kubenswrapper[4725]: I1202 13:17:15.555043 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:15 crc kubenswrapper[4725]: I1202 13:17:15.633546 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:16 crc kubenswrapper[4725]: I1202 13:17:16.254681 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:16 crc kubenswrapper[4725]: I1202 13:17:16.299298 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfj5"] Dec 02 13:17:18 crc kubenswrapper[4725]: I1202 13:17:18.196401 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kgfj5" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="registry-server" containerID="cri-o://12c1ef0c2cbd247d6f1c75cb6124975aed8a1b4459c1274e5203177df08a929d" gracePeriod=2 Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.237559 4725 generic.go:334] "Generic (PLEG): container finished" podID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerID="12c1ef0c2cbd247d6f1c75cb6124975aed8a1b4459c1274e5203177df08a929d" exitCode=0 Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.238319 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfj5" event={"ID":"ccfb8381-ea5f-4336-b453-0cb863e6ec96","Type":"ContainerDied","Data":"12c1ef0c2cbd247d6f1c75cb6124975aed8a1b4459c1274e5203177df08a929d"} Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.294646 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.361053 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-utilities\") pod \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.361087 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6v272\" (UniqueName: \"kubernetes.io/projected/ccfb8381-ea5f-4336-b453-0cb863e6ec96-kube-api-access-6v272\") pod \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.361109 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-catalog-content\") pod \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\" (UID: \"ccfb8381-ea5f-4336-b453-0cb863e6ec96\") " Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.361762 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-utilities" (OuterVolumeSpecName: "utilities") pod "ccfb8381-ea5f-4336-b453-0cb863e6ec96" (UID: "ccfb8381-ea5f-4336-b453-0cb863e6ec96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.367513 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccfb8381-ea5f-4336-b453-0cb863e6ec96-kube-api-access-6v272" (OuterVolumeSpecName: "kube-api-access-6v272") pod "ccfb8381-ea5f-4336-b453-0cb863e6ec96" (UID: "ccfb8381-ea5f-4336-b453-0cb863e6ec96"). InnerVolumeSpecName "kube-api-access-6v272". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.380261 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccfb8381-ea5f-4336-b453-0cb863e6ec96" (UID: "ccfb8381-ea5f-4336-b453-0cb863e6ec96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.462268 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.462317 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6v272\" (UniqueName: \"kubernetes.io/projected/ccfb8381-ea5f-4336-b453-0cb863e6ec96-kube-api-access-6v272\") on node \"crc\" DevicePath \"\"" Dec 02 13:17:19 crc kubenswrapper[4725]: I1202 13:17:19.462332 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfb8381-ea5f-4336-b453-0cb863e6ec96-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.244903 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgfj5" Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.244923 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgfj5" event={"ID":"ccfb8381-ea5f-4336-b453-0cb863e6ec96","Type":"ContainerDied","Data":"67dd0c081812dead9c800fabf7a0fbbc9692d36e38c88e00d21c683818031905"} Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.245373 4725 scope.go:117] "RemoveContainer" containerID="12c1ef0c2cbd247d6f1c75cb6124975aed8a1b4459c1274e5203177df08a929d" Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.247148 4725 generic.go:334] "Generic (PLEG): container finished" podID="46316422-5e47-461c-8827-6e12d896134c" containerID="5ca9af2c026456e3b32ac14636d5ab10ace1de93107bf86c8cc09db39cc51405" exitCode=0 Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.247221 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerDied","Data":"5ca9af2c026456e3b32ac14636d5ab10ace1de93107bf86c8cc09db39cc51405"} Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.251412 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" event={"ID":"00168d15-ce5e-4164-891c-1403031bee1d","Type":"ContainerStarted","Data":"f5ac5bb5b84ccb41cf39941fc890ff129f1e7c20b8eaadcf54a895431b03a243"} Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.251687 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.261231 4725 scope.go:117] "RemoveContainer" containerID="884a257ed1be387eba17185dd00a07b5ed1cf23cf38e495fd52d378b87b1b7f6" Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.287817 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfj5"] Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.292268 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgfj5"] Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.295423 4725 scope.go:117] "RemoveContainer" containerID="cf4984f621117916cc803b68337773004015a6933f5bb2a15a1c65cab46be75a" Dec 02 13:17:20 crc kubenswrapper[4725]: I1202 13:17:20.302330 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" podStartSLOduration=2.519739395 podStartE2EDuration="10.302313186s" podCreationTimestamp="2025-12-02 13:17:10 +0000 UTC" firstStartedPulling="2025-12-02 13:17:11.248325408 +0000 UTC m=+762.204967103" lastFinishedPulling="2025-12-02 13:17:19.030899199 +0000 UTC m=+769.987540894" observedRunningTime="2025-12-02 13:17:20.300276566 +0000 UTC m=+771.256918271" watchObservedRunningTime="2025-12-02 13:17:20.302313186 +0000 UTC m=+771.258954881" Dec 02 13:17:21 crc kubenswrapper[4725]: I1202 13:17:21.257607 4725 generic.go:334] "Generic (PLEG): container finished" podID="46316422-5e47-461c-8827-6e12d896134c" containerID="10b902823824f796396174ee2a45af4bd08bba0dfe7aaae383c00271720f553b" exitCode=0 Dec 02 13:17:21 crc kubenswrapper[4725]: I1202 13:17:21.257653 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerDied","Data":"10b902823824f796396174ee2a45af4bd08bba0dfe7aaae383c00271720f553b"} Dec 02 13:17:21 crc kubenswrapper[4725]: I1202 13:17:21.276952 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" path="/var/lib/kubelet/pods/ccfb8381-ea5f-4336-b453-0cb863e6ec96/volumes" Dec 02 13:17:22 crc kubenswrapper[4725]: I1202 13:17:22.266210 4725 generic.go:334] "Generic (PLEG): container finished" podID="46316422-5e47-461c-8827-6e12d896134c" containerID="df8e4c5d932c9e7f5f07471cec9e871d0fa6ba4237b603d7c3362afd6142ba7b" exitCode=0 Dec 02 13:17:22 crc kubenswrapper[4725]: I1202 13:17:22.266288 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerDied","Data":"df8e4c5d932c9e7f5f07471cec9e871d0fa6ba4237b603d7c3362afd6142ba7b"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276057 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"714c3092e39a943475987ca114ae3b2d59ae87db3b796bb1b1c67ec15c54b2cf"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276381 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276395 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"9b9cad34bb8ca4edffdd300fa175983cb39c97115e5670fff10e9249dc1f840b"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276405 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"eae7302764531742244f6cb32342753df08447be3f2cc8fadf0c7f201f34a2b4"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276415 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"e1446c94388a35e6dfa264d797f1f1f18c9a8dfaf150fb2accdde2dbea62dbf0"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276423 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"bdf8a52dd3982205d42716f13e6e18565c06468c21cf9bb4e2e6aa4088c1ef3c"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.276432 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-bspx6" event={"ID":"46316422-5e47-461c-8827-6e12d896134c","Type":"ContainerStarted","Data":"a7939ad007be37d3a218a940615505da01145aacfd185ba7835c11a2e1569ae3"} Dec 02 13:17:23 crc kubenswrapper[4725]: I1202 13:17:23.293904 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-bspx6" podStartSLOduration=5.379116445 podStartE2EDuration="13.293886268s" podCreationTimestamp="2025-12-02 13:17:10 +0000 UTC" firstStartedPulling="2025-12-02 13:17:11.139669492 +0000 UTC m=+762.096311187" lastFinishedPulling="2025-12-02 13:17:19.054439315 +0000 UTC m=+770.011081010" observedRunningTime="2025-12-02 13:17:23.292453872 +0000 UTC m=+774.249095557" watchObservedRunningTime="2025-12-02 13:17:23.293886268 +0000 UTC m=+774.250527963" Dec 02 13:17:26 crc kubenswrapper[4725]: I1202 13:17:26.004993 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:26 crc kubenswrapper[4725]: I1202 13:17:26.052863 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:30 crc kubenswrapper[4725]: I1202 13:17:30.520804 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-bkrzn" Dec 02 13:17:31 crc kubenswrapper[4725]: I1202 13:17:31.015980 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-c7g46" Dec 02 13:17:32 crc kubenswrapper[4725]: I1202 13:17:32.004721 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-dhmrd" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.071173 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2dldk"] Dec 02 13:17:35 crc kubenswrapper[4725]: E1202 13:17:35.072650 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="registry-server" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.072764 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="registry-server" Dec 02 13:17:35 crc kubenswrapper[4725]: E1202 13:17:35.072865 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="extract-content" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.072936 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="extract-content" Dec 02 13:17:35 crc kubenswrapper[4725]: E1202 13:17:35.073008 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="extract-utilities" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.073075 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="extract-utilities" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.073296 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccfb8381-ea5f-4336-b453-0cb863e6ec96" containerName="registry-server" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.073963 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.077853 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.078060 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.078165 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-9ppxp" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.084955 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2dldk"] Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.233725 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz8c6\" (UniqueName: \"kubernetes.io/projected/4493e01d-4f86-408e-92e3-d05bda19a5af-kube-api-access-vz8c6\") pod \"openstack-operator-index-2dldk\" (UID: \"4493e01d-4f86-408e-92e3-d05bda19a5af\") " pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.335711 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz8c6\" (UniqueName: \"kubernetes.io/projected/4493e01d-4f86-408e-92e3-d05bda19a5af-kube-api-access-vz8c6\") pod \"openstack-operator-index-2dldk\" (UID: \"4493e01d-4f86-408e-92e3-d05bda19a5af\") " pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.356939 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz8c6\" (UniqueName: \"kubernetes.io/projected/4493e01d-4f86-408e-92e3-d05bda19a5af-kube-api-access-vz8c6\") pod \"openstack-operator-index-2dldk\" (UID: \"4493e01d-4f86-408e-92e3-d05bda19a5af\") " pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.437183 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:35 crc kubenswrapper[4725]: I1202 13:17:35.804028 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2dldk"] Dec 02 13:17:35 crc kubenswrapper[4725]: W1202 13:17:35.809842 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4493e01d_4f86_408e_92e3_d05bda19a5af.slice/crio-dd09a1340fb792df0c7aa6f8af7333a63e251a6252c6cdeb55ad0f3c0e99ec71 WatchSource:0}: Error finding container dd09a1340fb792df0c7aa6f8af7333a63e251a6252c6cdeb55ad0f3c0e99ec71: Status 404 returned error can't find the container with id dd09a1340fb792df0c7aa6f8af7333a63e251a6252c6cdeb55ad0f3c0e99ec71 Dec 02 13:17:36 crc kubenswrapper[4725]: I1202 13:17:36.361409 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2dldk" event={"ID":"4493e01d-4f86-408e-92e3-d05bda19a5af","Type":"ContainerStarted","Data":"dd09a1340fb792df0c7aa6f8af7333a63e251a6252c6cdeb55ad0f3c0e99ec71"} Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.374404 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2dldk" event={"ID":"4493e01d-4f86-408e-92e3-d05bda19a5af","Type":"ContainerStarted","Data":"92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae"} Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.389280 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2dldk" podStartSLOduration=1.6401456859999999 podStartE2EDuration="3.38925901s" podCreationTimestamp="2025-12-02 13:17:35 +0000 UTC" firstStartedPulling="2025-12-02 13:17:35.812911031 +0000 UTC m=+786.769552726" lastFinishedPulling="2025-12-02 13:17:37.562024335 +0000 UTC m=+788.518666050" observedRunningTime="2025-12-02 13:17:38.386265185 +0000 UTC m=+789.342906880" watchObservedRunningTime="2025-12-02 13:17:38.38925901 +0000 UTC m=+789.345900705" Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.438489 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2dldk"] Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.900081 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-59k9r"] Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.901129 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.908397 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-59k9r"] Dec 02 13:17:38 crc kubenswrapper[4725]: I1202 13:17:38.998339 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcjt4\" (UniqueName: \"kubernetes.io/projected/6f802fa6-11b4-40ef-b9f1-bfb654645474-kube-api-access-jcjt4\") pod \"openstack-operator-index-59k9r\" (UID: \"6f802fa6-11b4-40ef-b9f1-bfb654645474\") " pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:39 crc kubenswrapper[4725]: I1202 13:17:39.099976 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcjt4\" (UniqueName: \"kubernetes.io/projected/6f802fa6-11b4-40ef-b9f1-bfb654645474-kube-api-access-jcjt4\") pod \"openstack-operator-index-59k9r\" (UID: \"6f802fa6-11b4-40ef-b9f1-bfb654645474\") " pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:39 crc kubenswrapper[4725]: I1202 13:17:39.117780 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcjt4\" (UniqueName: \"kubernetes.io/projected/6f802fa6-11b4-40ef-b9f1-bfb654645474-kube-api-access-jcjt4\") pod \"openstack-operator-index-59k9r\" (UID: \"6f802fa6-11b4-40ef-b9f1-bfb654645474\") " pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:39 crc kubenswrapper[4725]: I1202 13:17:39.221685 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:39 crc kubenswrapper[4725]: I1202 13:17:39.611917 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-59k9r"] Dec 02 13:17:39 crc kubenswrapper[4725]: W1202 13:17:39.616922 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f802fa6_11b4_40ef_b9f1_bfb654645474.slice/crio-5dca93ffb455fcdad96f8fa3cf9829a51789c8fd72fee01e082b073292e197ec WatchSource:0}: Error finding container 5dca93ffb455fcdad96f8fa3cf9829a51789c8fd72fee01e082b073292e197ec: Status 404 returned error can't find the container with id 5dca93ffb455fcdad96f8fa3cf9829a51789c8fd72fee01e082b073292e197ec Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.387975 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-59k9r" event={"ID":"6f802fa6-11b4-40ef-b9f1-bfb654645474","Type":"ContainerStarted","Data":"05e65db605c7f04851545061a4e0b87f6cb8aa56af4664cc5da5a4faea07a345"} Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.388280 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-59k9r" event={"ID":"6f802fa6-11b4-40ef-b9f1-bfb654645474","Type":"ContainerStarted","Data":"5dca93ffb455fcdad96f8fa3cf9829a51789c8fd72fee01e082b073292e197ec"} Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.388069 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2dldk" podUID="4493e01d-4f86-408e-92e3-d05bda19a5af" containerName="registry-server" containerID="cri-o://92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae" gracePeriod=2 Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.766572 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.783946 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-59k9r" podStartSLOduration=2.723905379 podStartE2EDuration="2.783930524s" podCreationTimestamp="2025-12-02 13:17:38 +0000 UTC" firstStartedPulling="2025-12-02 13:17:39.620976549 +0000 UTC m=+790.577618244" lastFinishedPulling="2025-12-02 13:17:39.681001694 +0000 UTC m=+790.637643389" observedRunningTime="2025-12-02 13:17:40.406810841 +0000 UTC m=+791.363452546" watchObservedRunningTime="2025-12-02 13:17:40.783930524 +0000 UTC m=+791.740572219" Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.925160 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz8c6\" (UniqueName: \"kubernetes.io/projected/4493e01d-4f86-408e-92e3-d05bda19a5af-kube-api-access-vz8c6\") pod \"4493e01d-4f86-408e-92e3-d05bda19a5af\" (UID: \"4493e01d-4f86-408e-92e3-d05bda19a5af\") " Dec 02 13:17:40 crc kubenswrapper[4725]: I1202 13:17:40.930374 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4493e01d-4f86-408e-92e3-d05bda19a5af-kube-api-access-vz8c6" (OuterVolumeSpecName: "kube-api-access-vz8c6") pod "4493e01d-4f86-408e-92e3-d05bda19a5af" (UID: "4493e01d-4f86-408e-92e3-d05bda19a5af"). InnerVolumeSpecName "kube-api-access-vz8c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.008129 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-bspx6" Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.026945 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz8c6\" (UniqueName: \"kubernetes.io/projected/4493e01d-4f86-408e-92e3-d05bda19a5af-kube-api-access-vz8c6\") on node \"crc\" DevicePath \"\"" Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.395097 4725 generic.go:334] "Generic (PLEG): container finished" podID="4493e01d-4f86-408e-92e3-d05bda19a5af" containerID="92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae" exitCode=0 Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.395142 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2dldk" Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.395158 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2dldk" event={"ID":"4493e01d-4f86-408e-92e3-d05bda19a5af","Type":"ContainerDied","Data":"92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae"} Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.395454 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2dldk" event={"ID":"4493e01d-4f86-408e-92e3-d05bda19a5af","Type":"ContainerDied","Data":"dd09a1340fb792df0c7aa6f8af7333a63e251a6252c6cdeb55ad0f3c0e99ec71"} Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.395493 4725 scope.go:117] "RemoveContainer" containerID="92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae" Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.412978 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2dldk"] Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.416357 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2dldk"] Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.417327 4725 scope.go:117] "RemoveContainer" containerID="92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae" Dec 02 13:17:41 crc kubenswrapper[4725]: E1202 13:17:41.417841 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae\": container with ID starting with 92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae not found: ID does not exist" containerID="92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae" Dec 02 13:17:41 crc kubenswrapper[4725]: I1202 13:17:41.417868 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae"} err="failed to get container status \"92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae\": rpc error: code = NotFound desc = could not find container \"92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae\": container with ID starting with 92eba61057fde37d2038b6748f51425ca459c0f854e055ab2ab76e2368ea0aae not found: ID does not exist" Dec 02 13:17:43 crc kubenswrapper[4725]: I1202 13:17:43.276110 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4493e01d-4f86-408e-92e3-d05bda19a5af" path="/var/lib/kubelet/pods/4493e01d-4f86-408e-92e3-d05bda19a5af/volumes" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.663933 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h84tg"] Dec 02 13:17:48 crc kubenswrapper[4725]: E1202 13:17:48.664292 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4493e01d-4f86-408e-92e3-d05bda19a5af" containerName="registry-server" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.664308 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="4493e01d-4f86-408e-92e3-d05bda19a5af" containerName="registry-server" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.664450 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="4493e01d-4f86-408e-92e3-d05bda19a5af" containerName="registry-server" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.665480 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.673095 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h84tg"] Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.844134 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs4nt\" (UniqueName: \"kubernetes.io/projected/2bada98b-49dd-4a21-8967-42e9b59310bd-kube-api-access-bs4nt\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.844186 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-catalog-content\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.844274 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-utilities\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.945215 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs4nt\" (UniqueName: \"kubernetes.io/projected/2bada98b-49dd-4a21-8967-42e9b59310bd-kube-api-access-bs4nt\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.945261 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-catalog-content\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.945322 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-utilities\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.945752 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-utilities\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.945974 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-catalog-content\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.966368 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs4nt\" (UniqueName: \"kubernetes.io/projected/2bada98b-49dd-4a21-8967-42e9b59310bd-kube-api-access-bs4nt\") pod \"redhat-operators-h84tg\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:48 crc kubenswrapper[4725]: I1202 13:17:48.982929 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:49 crc kubenswrapper[4725]: I1202 13:17:49.222613 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:49 crc kubenswrapper[4725]: I1202 13:17:49.222980 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:49 crc kubenswrapper[4725]: I1202 13:17:49.252789 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:49 crc kubenswrapper[4725]: I1202 13:17:49.421487 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h84tg"] Dec 02 13:17:49 crc kubenswrapper[4725]: I1202 13:17:49.445453 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerStarted","Data":"7522ab9833e5f2371887b9892e150a35370e1f8ffe2be9d14d937b85436eaba9"} Dec 02 13:17:49 crc kubenswrapper[4725]: I1202 13:17:49.469642 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-59k9r" Dec 02 13:17:50 crc kubenswrapper[4725]: I1202 13:17:50.452289 4725 generic.go:334] "Generic (PLEG): container finished" podID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerID="0464ef61a5c4d12ea2584f414c070c4d9c332276774fa63e673e4759dabb7ba6" exitCode=0 Dec 02 13:17:50 crc kubenswrapper[4725]: I1202 13:17:50.452349 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerDied","Data":"0464ef61a5c4d12ea2584f414c070c4d9c332276774fa63e673e4759dabb7ba6"} Dec 02 13:17:52 crc kubenswrapper[4725]: I1202 13:17:52.472546 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerStarted","Data":"28920892491d817310e089f5399a0e4f8e4dc07a9ff8e33076c0ccd322ed4c92"} Dec 02 13:17:54 crc kubenswrapper[4725]: I1202 13:17:54.485478 4725 generic.go:334] "Generic (PLEG): container finished" podID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerID="28920892491d817310e089f5399a0e4f8e4dc07a9ff8e33076c0ccd322ed4c92" exitCode=0 Dec 02 13:17:54 crc kubenswrapper[4725]: I1202 13:17:54.485594 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerDied","Data":"28920892491d817310e089f5399a0e4f8e4dc07a9ff8e33076c0ccd322ed4c92"} Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.490048 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745"] Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.491481 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.493331 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerStarted","Data":"ea574503cb8f3b73da236c3d39de6c1b04292c779b987eba99d179854a70370d"} Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.495105 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xxbv7" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.500002 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745"] Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.592770 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h84tg" podStartSLOduration=3.090566302 podStartE2EDuration="7.592745107s" podCreationTimestamp="2025-12-02 13:17:48 +0000 UTC" firstStartedPulling="2025-12-02 13:17:50.45452278 +0000 UTC m=+801.411164505" lastFinishedPulling="2025-12-02 13:17:54.956701615 +0000 UTC m=+805.913343310" observedRunningTime="2025-12-02 13:17:55.578985804 +0000 UTC m=+806.535627519" watchObservedRunningTime="2025-12-02 13:17:55.592745107 +0000 UTC m=+806.549386802" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.656866 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-bundle\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.656941 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vbk2\" (UniqueName: \"kubernetes.io/projected/856ce13a-68af-4c76-938a-12df01458fa3-kube-api-access-2vbk2\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.657032 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-util\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.759654 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-util\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.759720 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-bundle\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.759748 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vbk2\" (UniqueName: \"kubernetes.io/projected/856ce13a-68af-4c76-938a-12df01458fa3-kube-api-access-2vbk2\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.760377 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-bundle\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.760669 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-util\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.788232 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vbk2\" (UniqueName: \"kubernetes.io/projected/856ce13a-68af-4c76-938a-12df01458fa3-kube-api-access-2vbk2\") pod \"0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:55 crc kubenswrapper[4725]: I1202 13:17:55.867213 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:17:56 crc kubenswrapper[4725]: I1202 13:17:56.087146 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745"] Dec 02 13:17:56 crc kubenswrapper[4725]: I1202 13:17:56.500181 4725 generic.go:334] "Generic (PLEG): container finished" podID="856ce13a-68af-4c76-938a-12df01458fa3" containerID="bc4fdea37f83dea6da69d370a9603ab24cc550048d23d5b91ed2e37dec1841c7" exitCode=0 Dec 02 13:17:56 crc kubenswrapper[4725]: I1202 13:17:56.500241 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" event={"ID":"856ce13a-68af-4c76-938a-12df01458fa3","Type":"ContainerDied","Data":"bc4fdea37f83dea6da69d370a9603ab24cc550048d23d5b91ed2e37dec1841c7"} Dec 02 13:17:56 crc kubenswrapper[4725]: I1202 13:17:56.500278 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" event={"ID":"856ce13a-68af-4c76-938a-12df01458fa3","Type":"ContainerStarted","Data":"99e4696117711b31b916b11d7f25ae6ba3c9bbafebd426913a0379e9ee0ce4bf"} Dec 02 13:17:57 crc kubenswrapper[4725]: I1202 13:17:57.508040 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" event={"ID":"856ce13a-68af-4c76-938a-12df01458fa3","Type":"ContainerStarted","Data":"09049672a05a6c54bb0dcbb7cc6d7fe96545e413c153eedc55988fdd25983cf4"} Dec 02 13:17:58 crc kubenswrapper[4725]: I1202 13:17:58.515795 4725 generic.go:334] "Generic (PLEG): container finished" podID="856ce13a-68af-4c76-938a-12df01458fa3" containerID="09049672a05a6c54bb0dcbb7cc6d7fe96545e413c153eedc55988fdd25983cf4" exitCode=0 Dec 02 13:17:58 crc kubenswrapper[4725]: I1202 13:17:58.515842 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" event={"ID":"856ce13a-68af-4c76-938a-12df01458fa3","Type":"ContainerDied","Data":"09049672a05a6c54bb0dcbb7cc6d7fe96545e413c153eedc55988fdd25983cf4"} Dec 02 13:17:58 crc kubenswrapper[4725]: I1202 13:17:58.983663 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:58 crc kubenswrapper[4725]: I1202 13:17:58.984013 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:17:59 crc kubenswrapper[4725]: I1202 13:17:59.523485 4725 generic.go:334] "Generic (PLEG): container finished" podID="856ce13a-68af-4c76-938a-12df01458fa3" containerID="e2e33604663a9673ff23d2b29df3e4f0f6fb6bb924afaf221a81d5cb85c9234c" exitCode=0 Dec 02 13:17:59 crc kubenswrapper[4725]: I1202 13:17:59.523524 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" event={"ID":"856ce13a-68af-4c76-938a-12df01458fa3","Type":"ContainerDied","Data":"e2e33604663a9673ff23d2b29df3e4f0f6fb6bb924afaf221a81d5cb85c9234c"} Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.026632 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h84tg" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="registry-server" probeResult="failure" output=< Dec 02 13:18:00 crc kubenswrapper[4725]: timeout: failed to connect service ":50051" within 1s Dec 02 13:18:00 crc kubenswrapper[4725]: > Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.761411 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.931769 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vbk2\" (UniqueName: \"kubernetes.io/projected/856ce13a-68af-4c76-938a-12df01458fa3-kube-api-access-2vbk2\") pod \"856ce13a-68af-4c76-938a-12df01458fa3\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.931847 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-util\") pod \"856ce13a-68af-4c76-938a-12df01458fa3\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.931894 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-bundle\") pod \"856ce13a-68af-4c76-938a-12df01458fa3\" (UID: \"856ce13a-68af-4c76-938a-12df01458fa3\") " Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.932645 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-bundle" (OuterVolumeSpecName: "bundle") pod "856ce13a-68af-4c76-938a-12df01458fa3" (UID: "856ce13a-68af-4c76-938a-12df01458fa3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.937674 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/856ce13a-68af-4c76-938a-12df01458fa3-kube-api-access-2vbk2" (OuterVolumeSpecName: "kube-api-access-2vbk2") pod "856ce13a-68af-4c76-938a-12df01458fa3" (UID: "856ce13a-68af-4c76-938a-12df01458fa3"). InnerVolumeSpecName "kube-api-access-2vbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:18:00 crc kubenswrapper[4725]: I1202 13:18:00.945998 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-util" (OuterVolumeSpecName: "util") pod "856ce13a-68af-4c76-938a-12df01458fa3" (UID: "856ce13a-68af-4c76-938a-12df01458fa3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:18:01 crc kubenswrapper[4725]: I1202 13:18:01.033082 4725 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:18:01 crc kubenswrapper[4725]: I1202 13:18:01.033124 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vbk2\" (UniqueName: \"kubernetes.io/projected/856ce13a-68af-4c76-938a-12df01458fa3-kube-api-access-2vbk2\") on node \"crc\" DevicePath \"\"" Dec 02 13:18:01 crc kubenswrapper[4725]: I1202 13:18:01.033140 4725 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/856ce13a-68af-4c76-938a-12df01458fa3-util\") on node \"crc\" DevicePath \"\"" Dec 02 13:18:01 crc kubenswrapper[4725]: I1202 13:18:01.536165 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" event={"ID":"856ce13a-68af-4c76-938a-12df01458fa3","Type":"ContainerDied","Data":"99e4696117711b31b916b11d7f25ae6ba3c9bbafebd426913a0379e9ee0ce4bf"} Dec 02 13:18:01 crc kubenswrapper[4725]: I1202 13:18:01.536210 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99e4696117711b31b916b11d7f25ae6ba3c9bbafebd426913a0379e9ee0ce4bf" Dec 02 13:18:01 crc kubenswrapper[4725]: I1202 13:18:01.536237 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.262021 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf"] Dec 02 13:18:05 crc kubenswrapper[4725]: E1202 13:18:05.263902 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="pull" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.264028 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="pull" Dec 02 13:18:05 crc kubenswrapper[4725]: E1202 13:18:05.264108 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="util" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.264183 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="util" Dec 02 13:18:05 crc kubenswrapper[4725]: E1202 13:18:05.264268 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="extract" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.264338 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="extract" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.264580 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="856ce13a-68af-4c76-938a-12df01458fa3" containerName="extract" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.265200 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.267864 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-nl7s4" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.301717 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf"] Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.385912 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt7jv\" (UniqueName: \"kubernetes.io/projected/98fc37a5-6d29-4f0b-93ab-ffecc2157b33-kube-api-access-lt7jv\") pod \"openstack-operator-controller-operator-75ff9f7b7-wcxkf\" (UID: \"98fc37a5-6d29-4f0b-93ab-ffecc2157b33\") " pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.488283 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt7jv\" (UniqueName: \"kubernetes.io/projected/98fc37a5-6d29-4f0b-93ab-ffecc2157b33-kube-api-access-lt7jv\") pod \"openstack-operator-controller-operator-75ff9f7b7-wcxkf\" (UID: \"98fc37a5-6d29-4f0b-93ab-ffecc2157b33\") " pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.512682 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt7jv\" (UniqueName: \"kubernetes.io/projected/98fc37a5-6d29-4f0b-93ab-ffecc2157b33-kube-api-access-lt7jv\") pod \"openstack-operator-controller-operator-75ff9f7b7-wcxkf\" (UID: \"98fc37a5-6d29-4f0b-93ab-ffecc2157b33\") " pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.582251 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:05 crc kubenswrapper[4725]: I1202 13:18:05.821050 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf"] Dec 02 13:18:06 crc kubenswrapper[4725]: I1202 13:18:06.574299 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" event={"ID":"98fc37a5-6d29-4f0b-93ab-ffecc2157b33","Type":"ContainerStarted","Data":"92a089ceb17dbfd23fc8248377d244de95bc6f49d90e059b3304d194b120cbbc"} Dec 02 13:18:09 crc kubenswrapper[4725]: I1202 13:18:09.063585 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:18:09 crc kubenswrapper[4725]: I1202 13:18:09.110725 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:18:10 crc kubenswrapper[4725]: I1202 13:18:10.600028 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" event={"ID":"98fc37a5-6d29-4f0b-93ab-ffecc2157b33","Type":"ContainerStarted","Data":"f3be81d403692df549807033993c8c1865016578896d2fd55e911d4d8083e770"} Dec 02 13:18:10 crc kubenswrapper[4725]: I1202 13:18:10.600162 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:10 crc kubenswrapper[4725]: I1202 13:18:10.635691 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" podStartSLOduration=1.815991275 podStartE2EDuration="5.635672183s" podCreationTimestamp="2025-12-02 13:18:05 +0000 UTC" firstStartedPulling="2025-12-02 13:18:05.831288799 +0000 UTC m=+816.787930494" lastFinishedPulling="2025-12-02 13:18:09.650969687 +0000 UTC m=+820.607611402" observedRunningTime="2025-12-02 13:18:10.631563391 +0000 UTC m=+821.588205096" watchObservedRunningTime="2025-12-02 13:18:10.635672183 +0000 UTC m=+821.592313878" Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.442190 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h84tg"] Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.442662 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h84tg" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="registry-server" containerID="cri-o://ea574503cb8f3b73da236c3d39de6c1b04292c779b987eba99d179854a70370d" gracePeriod=2 Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.606639 4725 generic.go:334] "Generic (PLEG): container finished" podID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerID="ea574503cb8f3b73da236c3d39de6c1b04292c779b987eba99d179854a70370d" exitCode=0 Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.606708 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerDied","Data":"ea574503cb8f3b73da236c3d39de6c1b04292c779b987eba99d179854a70370d"} Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.760106 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.872943 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-catalog-content\") pod \"2bada98b-49dd-4a21-8967-42e9b59310bd\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.873274 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-utilities\") pod \"2bada98b-49dd-4a21-8967-42e9b59310bd\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.873325 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs4nt\" (UniqueName: \"kubernetes.io/projected/2bada98b-49dd-4a21-8967-42e9b59310bd-kube-api-access-bs4nt\") pod \"2bada98b-49dd-4a21-8967-42e9b59310bd\" (UID: \"2bada98b-49dd-4a21-8967-42e9b59310bd\") " Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.874115 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-utilities" (OuterVolumeSpecName: "utilities") pod "2bada98b-49dd-4a21-8967-42e9b59310bd" (UID: "2bada98b-49dd-4a21-8967-42e9b59310bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.878647 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bada98b-49dd-4a21-8967-42e9b59310bd-kube-api-access-bs4nt" (OuterVolumeSpecName: "kube-api-access-bs4nt") pod "2bada98b-49dd-4a21-8967-42e9b59310bd" (UID: "2bada98b-49dd-4a21-8967-42e9b59310bd"). InnerVolumeSpecName "kube-api-access-bs4nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.975270 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.975310 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs4nt\" (UniqueName: \"kubernetes.io/projected/2bada98b-49dd-4a21-8967-42e9b59310bd-kube-api-access-bs4nt\") on node \"crc\" DevicePath \"\"" Dec 02 13:18:11 crc kubenswrapper[4725]: I1202 13:18:11.975845 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bada98b-49dd-4a21-8967-42e9b59310bd" (UID: "2bada98b-49dd-4a21-8967-42e9b59310bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.077104 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bada98b-49dd-4a21-8967-42e9b59310bd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.612752 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h84tg" event={"ID":"2bada98b-49dd-4a21-8967-42e9b59310bd","Type":"ContainerDied","Data":"7522ab9833e5f2371887b9892e150a35370e1f8ffe2be9d14d937b85436eaba9"} Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.612809 4725 scope.go:117] "RemoveContainer" containerID="ea574503cb8f3b73da236c3d39de6c1b04292c779b987eba99d179854a70370d" Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.612855 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h84tg" Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.628044 4725 scope.go:117] "RemoveContainer" containerID="28920892491d817310e089f5399a0e4f8e4dc07a9ff8e33076c0ccd322ed4c92" Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.640983 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h84tg"] Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.647180 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h84tg"] Dec 02 13:18:12 crc kubenswrapper[4725]: I1202 13:18:12.649496 4725 scope.go:117] "RemoveContainer" containerID="0464ef61a5c4d12ea2584f414c070c4d9c332276774fa63e673e4759dabb7ba6" Dec 02 13:18:13 crc kubenswrapper[4725]: I1202 13:18:13.277259 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" path="/var/lib/kubelet/pods/2bada98b-49dd-4a21-8967-42e9b59310bd/volumes" Dec 02 13:18:15 crc kubenswrapper[4725]: I1202 13:18:15.585760 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-75ff9f7b7-wcxkf" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.955283 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq"] Dec 02 13:18:34 crc kubenswrapper[4725]: E1202 13:18:34.956008 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="registry-server" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.956020 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="registry-server" Dec 02 13:18:34 crc kubenswrapper[4725]: E1202 13:18:34.956031 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="extract-content" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.956036 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="extract-content" Dec 02 13:18:34 crc kubenswrapper[4725]: E1202 13:18:34.956057 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="extract-utilities" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.956064 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="extract-utilities" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.956164 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bada98b-49dd-4a21-8967-42e9b59310bd" containerName="registry-server" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.956749 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.959489 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-bdt6q" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.967558 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg"] Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.968624 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.972121 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-8pl7r" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.972622 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2xpk\" (UniqueName: \"kubernetes.io/projected/b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b-kube-api-access-t2xpk\") pod \"barbican-operator-controller-manager-7d9dfd778-ctzjq\" (UID: \"b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.972697 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh5rt\" (UniqueName: \"kubernetes.io/projected/570a3a73-dffb-4ded-8fd1-d30ddc60979d-kube-api-access-wh5rt\") pod \"cinder-operator-controller-manager-859b6ccc6-z9ttg\" (UID: \"570a3a73-dffb-4ded-8fd1-d30ddc60979d\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:18:34 crc kubenswrapper[4725]: I1202 13:18:34.980697 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.021636 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.022611 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.027269 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.027373 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-z2v9s" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.045514 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.046510 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.046595 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.054162 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-hlfz7" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.054348 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.055271 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.060048 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-52w2x" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.065003 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.075072 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.075269 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dnx5\" (UniqueName: \"kubernetes.io/projected/c42118d3-322d-4733-8749-d13bc174a6f8-kube-api-access-9dnx5\") pod \"designate-operator-controller-manager-78b4bc895b-ppgnw\" (UID: \"c42118d3-322d-4733-8749-d13bc174a6f8\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.075308 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz2v6\" (UniqueName: \"kubernetes.io/projected/0b562f2a-162c-4611-81d3-00937e945f96-kube-api-access-zz2v6\") pod \"heat-operator-controller-manager-5f64f6f8bb-cwfk4\" (UID: \"0b562f2a-162c-4611-81d3-00937e945f96\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.075362 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2xpk\" (UniqueName: \"kubernetes.io/projected/b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b-kube-api-access-t2xpk\") pod \"barbican-operator-controller-manager-7d9dfd778-ctzjq\" (UID: \"b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.075393 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh5rt\" (UniqueName: \"kubernetes.io/projected/570a3a73-dffb-4ded-8fd1-d30ddc60979d-kube-api-access-wh5rt\") pod \"cinder-operator-controller-manager-859b6ccc6-z9ttg\" (UID: \"570a3a73-dffb-4ded-8fd1-d30ddc60979d\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.075426 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgsrc\" (UniqueName: \"kubernetes.io/projected/65a9f532-387a-487e-bfe0-81a110509366-kube-api-access-pgsrc\") pod \"glance-operator-controller-manager-77987cd8cd-njlj7\" (UID: \"65a9f532-387a-487e-bfe0-81a110509366\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.113342 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2xpk\" (UniqueName: \"kubernetes.io/projected/b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b-kube-api-access-t2xpk\") pod \"barbican-operator-controller-manager-7d9dfd778-ctzjq\" (UID: \"b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.114160 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh5rt\" (UniqueName: \"kubernetes.io/projected/570a3a73-dffb-4ded-8fd1-d30ddc60979d-kube-api-access-wh5rt\") pod \"cinder-operator-controller-manager-859b6ccc6-z9ttg\" (UID: \"570a3a73-dffb-4ded-8fd1-d30ddc60979d\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.138525 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.139895 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.148170 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-p9pdv" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.156194 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g99db"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.157189 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.162494 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.162699 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-lb7sj" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.169348 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.170229 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.176168 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.177025 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dnx5\" (UniqueName: \"kubernetes.io/projected/c42118d3-322d-4733-8749-d13bc174a6f8-kube-api-access-9dnx5\") pod \"designate-operator-controller-manager-78b4bc895b-ppgnw\" (UID: \"c42118d3-322d-4733-8749-d13bc174a6f8\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.177060 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz2v6\" (UniqueName: \"kubernetes.io/projected/0b562f2a-162c-4611-81d3-00937e945f96-kube-api-access-zz2v6\") pod \"heat-operator-controller-manager-5f64f6f8bb-cwfk4\" (UID: \"0b562f2a-162c-4611-81d3-00937e945f96\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.177126 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgsrc\" (UniqueName: \"kubernetes.io/projected/65a9f532-387a-487e-bfe0-81a110509366-kube-api-access-pgsrc\") pod \"glance-operator-controller-manager-77987cd8cd-njlj7\" (UID: \"65a9f532-387a-487e-bfe0-81a110509366\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.177221 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ttc6x" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.180341 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g99db"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.254964 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.255553 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dnx5\" (UniqueName: \"kubernetes.io/projected/c42118d3-322d-4733-8749-d13bc174a6f8-kube-api-access-9dnx5\") pod \"designate-operator-controller-manager-78b4bc895b-ppgnw\" (UID: \"c42118d3-322d-4733-8749-d13bc174a6f8\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.256237 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgsrc\" (UniqueName: \"kubernetes.io/projected/65a9f532-387a-487e-bfe0-81a110509366-kube-api-access-pgsrc\") pod \"glance-operator-controller-manager-77987cd8cd-njlj7\" (UID: \"65a9f532-387a-487e-bfe0-81a110509366\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.272494 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz2v6\" (UniqueName: \"kubernetes.io/projected/0b562f2a-162c-4611-81d3-00937e945f96-kube-api-access-zz2v6\") pod \"heat-operator-controller-manager-5f64f6f8bb-cwfk4\" (UID: \"0b562f2a-162c-4611-81d3-00937e945f96\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.283712 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.285270 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpvtx\" (UniqueName: \"kubernetes.io/projected/4bad83be-e3ed-499a-8e64-cfe83c4ac1e6-kube-api-access-dpvtx\") pod \"ironic-operator-controller-manager-6c548fd776-klrxl\" (UID: \"4bad83be-e3ed-499a-8e64-cfe83c4ac1e6\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.285392 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkdqr\" (UniqueName: \"kubernetes.io/projected/6b08a8fe-08d5-4b04-9283-51973a9ac3d6-kube-api-access-pkdqr\") pod \"horizon-operator-controller-manager-68c6d99b8f-pmdw6\" (UID: \"6b08a8fe-08d5-4b04-9283-51973a9ac3d6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.285581 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.286399 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdvz8\" (UniqueName: \"kubernetes.io/projected/bf404326-3c43-4cc4-93b7-0793213afcba-kube-api-access-tdvz8\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.304196 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.351739 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.354260 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.359968 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.360805 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.361255 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.365906 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.371749 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-j82c6" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.372015 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-tmhl8" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.379720 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.389745 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkdqr\" (UniqueName: \"kubernetes.io/projected/6b08a8fe-08d5-4b04-9283-51973a9ac3d6-kube-api-access-pkdqr\") pod \"horizon-operator-controller-manager-68c6d99b8f-pmdw6\" (UID: \"6b08a8fe-08d5-4b04-9283-51973a9ac3d6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.389804 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.389835 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdvz8\" (UniqueName: \"kubernetes.io/projected/bf404326-3c43-4cc4-93b7-0793213afcba-kube-api-access-tdvz8\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.389862 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpvtx\" (UniqueName: \"kubernetes.io/projected/4bad83be-e3ed-499a-8e64-cfe83c4ac1e6-kube-api-access-dpvtx\") pod \"ironic-operator-controller-manager-6c548fd776-klrxl\" (UID: \"4bad83be-e3ed-499a-8e64-cfe83c4ac1e6\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.390287 4725 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.390335 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert podName:bf404326-3c43-4cc4-93b7-0793213afcba nodeName:}" failed. No retries permitted until 2025-12-02 13:18:35.890319959 +0000 UTC m=+846.846961644 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert") pod "infra-operator-controller-manager-57548d458d-g99db" (UID: "bf404326-3c43-4cc4-93b7-0793213afcba") : secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.391007 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.406103 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.422528 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.431000 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.436814 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.437972 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.439701 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-k9wws" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.451709 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.453933 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.455159 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.456830 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpvtx\" (UniqueName: \"kubernetes.io/projected/4bad83be-e3ed-499a-8e64-cfe83c4ac1e6-kube-api-access-dpvtx\") pod \"ironic-operator-controller-manager-6c548fd776-klrxl\" (UID: \"4bad83be-e3ed-499a-8e64-cfe83c4ac1e6\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.464148 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.464199 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.470761 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-h5jkd" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.471063 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-tp8zn" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.485041 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdvz8\" (UniqueName: \"kubernetes.io/projected/bf404326-3c43-4cc4-93b7-0793213afcba-kube-api-access-tdvz8\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.490624 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vwjn\" (UniqueName: \"kubernetes.io/projected/3fb7e977-cd14-4b71-b349-19d4487cfb15-kube-api-access-7vwjn\") pod \"keystone-operator-controller-manager-7765d96ddf-p22fj\" (UID: \"3fb7e977-cd14-4b71-b349-19d4487cfb15\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.490658 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdzmw\" (UniqueName: \"kubernetes.io/projected/c1f336f6-716b-4096-89cd-f23d1fed90cb-kube-api-access-fdzmw\") pod \"manila-operator-controller-manager-7c79b5df47-rk99q\" (UID: \"c1f336f6-716b-4096-89cd-f23d1fed90cb\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.500574 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkdqr\" (UniqueName: \"kubernetes.io/projected/6b08a8fe-08d5-4b04-9283-51973a9ac3d6-kube-api-access-pkdqr\") pod \"horizon-operator-controller-manager-68c6d99b8f-pmdw6\" (UID: \"6b08a8fe-08d5-4b04-9283-51973a9ac3d6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.532331 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-zdj64"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.533313 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.551170 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-bznkp" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.607898 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szh7z\" (UniqueName: \"kubernetes.io/projected/bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b-kube-api-access-szh7z\") pod \"octavia-operator-controller-manager-998648c74-zdj64\" (UID: \"bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.608157 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46tfq\" (UniqueName: \"kubernetes.io/projected/2c988ae2-9613-4584-a599-a7d063611189-kube-api-access-46tfq\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-wd49v\" (UID: \"2c988ae2-9613-4584-a599-a7d063611189\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.608178 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltc9k\" (UniqueName: \"kubernetes.io/projected/eacfb091-583b-4e89-9850-cc146cef7e20-kube-api-access-ltc9k\") pod \"nova-operator-controller-manager-697bc559fc-cllpp\" (UID: \"eacfb091-583b-4e89-9850-cc146cef7e20\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.608219 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vwjn\" (UniqueName: \"kubernetes.io/projected/3fb7e977-cd14-4b71-b349-19d4487cfb15-kube-api-access-7vwjn\") pod \"keystone-operator-controller-manager-7765d96ddf-p22fj\" (UID: \"3fb7e977-cd14-4b71-b349-19d4487cfb15\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.608237 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpfbt\" (UniqueName: \"kubernetes.io/projected/8f32e184-607a-4077-aa41-ca7aabba76aa-kube-api-access-qpfbt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-8s44s\" (UID: \"8f32e184-607a-4077-aa41-ca7aabba76aa\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.608257 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdzmw\" (UniqueName: \"kubernetes.io/projected/c1f336f6-716b-4096-89cd-f23d1fed90cb-kube-api-access-fdzmw\") pod \"manila-operator-controller-manager-7c79b5df47-rk99q\" (UID: \"c1f336f6-716b-4096-89cd-f23d1fed90cb\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.608545 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.612525 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.622638 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.657249 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-p7mdk" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.657792 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.672794 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-zdj64"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.694940 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.702094 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.703147 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.703483 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdzmw\" (UniqueName: \"kubernetes.io/projected/c1f336f6-716b-4096-89cd-f23d1fed90cb-kube-api-access-fdzmw\") pod \"manila-operator-controller-manager-7c79b5df47-rk99q\" (UID: \"c1f336f6-716b-4096-89cd-f23d1fed90cb\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:18:35 crc kubenswrapper[4725]: W1202 13:18:35.716233 4725 reflector.go:561] object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7w6j5": failed to list *v1.Secret: secrets "ovn-operator-controller-manager-dockercfg-7w6j5" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.716264 4725 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"ovn-operator-controller-manager-dockercfg-7w6j5\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-operator-controller-manager-dockercfg-7w6j5\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.717255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szh7z\" (UniqueName: \"kubernetes.io/projected/bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b-kube-api-access-szh7z\") pod \"octavia-operator-controller-manager-998648c74-zdj64\" (UID: \"bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.717298 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46tfq\" (UniqueName: \"kubernetes.io/projected/2c988ae2-9613-4584-a599-a7d063611189-kube-api-access-46tfq\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-wd49v\" (UID: \"2c988ae2-9613-4584-a599-a7d063611189\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.717318 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltc9k\" (UniqueName: \"kubernetes.io/projected/eacfb091-583b-4e89-9850-cc146cef7e20-kube-api-access-ltc9k\") pod \"nova-operator-controller-manager-697bc559fc-cllpp\" (UID: \"eacfb091-583b-4e89-9850-cc146cef7e20\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.717378 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpfbt\" (UniqueName: \"kubernetes.io/projected/8f32e184-607a-4077-aa41-ca7aabba76aa-kube-api-access-qpfbt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-8s44s\" (UID: \"8f32e184-607a-4077-aa41-ca7aabba76aa\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.760238 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpfbt\" (UniqueName: \"kubernetes.io/projected/8f32e184-607a-4077-aa41-ca7aabba76aa-kube-api-access-qpfbt\") pod \"mariadb-operator-controller-manager-56bbcc9d85-8s44s\" (UID: \"8f32e184-607a-4077-aa41-ca7aabba76aa\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.765452 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szh7z\" (UniqueName: \"kubernetes.io/projected/bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b-kube-api-access-szh7z\") pod \"octavia-operator-controller-manager-998648c74-zdj64\" (UID: \"bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.771257 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46tfq\" (UniqueName: \"kubernetes.io/projected/2c988ae2-9613-4584-a599-a7d063611189-kube-api-access-46tfq\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-wd49v\" (UID: \"2c988ae2-9613-4584-a599-a7d063611189\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.786324 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.788773 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltc9k\" (UniqueName: \"kubernetes.io/projected/eacfb091-583b-4e89-9850-cc146cef7e20-kube-api-access-ltc9k\") pod \"nova-operator-controller-manager-697bc559fc-cllpp\" (UID: \"eacfb091-583b-4e89-9850-cc146cef7e20\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.796029 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vwjn\" (UniqueName: \"kubernetes.io/projected/3fb7e977-cd14-4b71-b349-19d4487cfb15-kube-api-access-7vwjn\") pod \"keystone-operator-controller-manager-7765d96ddf-p22fj\" (UID: \"3fb7e977-cd14-4b71-b349-19d4487cfb15\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.811520 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.818358 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49x4t\" (UniqueName: \"kubernetes.io/projected/9c4065a8-50f2-4be9-bdd7-42345814f4c5-kube-api-access-49x4t\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.818422 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvmcm\" (UniqueName: \"kubernetes.io/projected/511bab3c-3885-4c5a-8c91-bd231b1c2a03-kube-api-access-jvmcm\") pod \"ovn-operator-controller-manager-b6456fdb6-mxpqd\" (UID: \"511bab3c-3885-4c5a-8c91-bd231b1c2a03\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.818498 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.827787 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.843579 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.854690 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.860354 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.871261 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-hb7c2" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.887598 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.889809 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.899361 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.900415 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.905807 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.906427 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-c578j" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.914933 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.915031 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.917064 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5pffq" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.920909 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49x4t\" (UniqueName: \"kubernetes.io/projected/9c4065a8-50f2-4be9-bdd7-42345814f4c5-kube-api-access-49x4t\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.920954 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvmcm\" (UniqueName: \"kubernetes.io/projected/511bab3c-3885-4c5a-8c91-bd231b1c2a03-kube-api-access-jvmcm\") pod \"ovn-operator-controller-manager-b6456fdb6-mxpqd\" (UID: \"511bab3c-3885-4c5a-8c91-bd231b1c2a03\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.921009 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqthf\" (UniqueName: \"kubernetes.io/projected/ad9c89d4-9792-401b-9f1f-81d2e5c7ede9-kube-api-access-pqthf\") pod \"placement-operator-controller-manager-78f8948974-wqn2j\" (UID: \"ad9c89d4-9792-401b-9f1f-81d2e5c7ede9\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.921046 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.921090 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.921121 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvp2j\" (UniqueName: \"kubernetes.io/projected/bd74c0bd-9ef5-487c-b190-2b23a8120c8d-kube-api-access-wvp2j\") pod \"telemetry-operator-controller-manager-76cc84c6bb-2l2x4\" (UID: \"bd74c0bd-9ef5-487c-b190-2b23a8120c8d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.921181 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwjb5\" (UniqueName: \"kubernetes.io/projected/dc849752-f58a-4e7a-b1fa-5b7d88c14d1a-kube-api-access-mwjb5\") pod \"swift-operator-controller-manager-5f8c65bbfc-cptsj\" (UID: \"dc849752-f58a-4e7a-b1fa-5b7d88c14d1a\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.921797 4725 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.921847 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert podName:9c4065a8-50f2-4be9-bdd7-42345814f4c5 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:36.421831307 +0000 UTC m=+847.378473002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" (UID: "9c4065a8-50f2-4be9-bdd7-42345814f4c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.926441 4725 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:35 crc kubenswrapper[4725]: E1202 13:18:35.926515 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert podName:bf404326-3c43-4cc4-93b7-0793213afcba nodeName:}" failed. No retries permitted until 2025-12-02 13:18:36.926500882 +0000 UTC m=+847.883142577 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert") pod "infra-operator-controller-manager-57548d458d-g99db" (UID: "bf404326-3c43-4cc4-93b7-0793213afcba") : secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.943699 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.943853 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4"] Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.949278 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49x4t\" (UniqueName: \"kubernetes.io/projected/9c4065a8-50f2-4be9-bdd7-42345814f4c5-kube-api-access-49x4t\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:35 crc kubenswrapper[4725]: I1202 13:18:35.949341 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.015233 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvmcm\" (UniqueName: \"kubernetes.io/projected/511bab3c-3885-4c5a-8c91-bd231b1c2a03-kube-api-access-jvmcm\") pod \"ovn-operator-controller-manager-b6456fdb6-mxpqd\" (UID: \"511bab3c-3885-4c5a-8c91-bd231b1c2a03\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.017175 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.029965 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.031924 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7p42c" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.044494 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvp2j\" (UniqueName: \"kubernetes.io/projected/bd74c0bd-9ef5-487c-b190-2b23a8120c8d-kube-api-access-wvp2j\") pod \"telemetry-operator-controller-manager-76cc84c6bb-2l2x4\" (UID: \"bd74c0bd-9ef5-487c-b190-2b23a8120c8d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.044665 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwjb5\" (UniqueName: \"kubernetes.io/projected/dc849752-f58a-4e7a-b1fa-5b7d88c14d1a-kube-api-access-mwjb5\") pod \"swift-operator-controller-manager-5f8c65bbfc-cptsj\" (UID: \"dc849752-f58a-4e7a-b1fa-5b7d88c14d1a\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.045398 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqthf\" (UniqueName: \"kubernetes.io/projected/ad9c89d4-9792-401b-9f1f-81d2e5c7ede9-kube-api-access-pqthf\") pod \"placement-operator-controller-manager-78f8948974-wqn2j\" (UID: \"ad9c89d4-9792-401b-9f1f-81d2e5c7ede9\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.049117 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.073774 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.075439 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.092568 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9ws6v" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.094513 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.098289 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvp2j\" (UniqueName: \"kubernetes.io/projected/bd74c0bd-9ef5-487c-b190-2b23a8120c8d-kube-api-access-wvp2j\") pod \"telemetry-operator-controller-manager-76cc84c6bb-2l2x4\" (UID: \"bd74c0bd-9ef5-487c-b190-2b23a8120c8d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.099257 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwjb5\" (UniqueName: \"kubernetes.io/projected/dc849752-f58a-4e7a-b1fa-5b7d88c14d1a-kube-api-access-mwjb5\") pod \"swift-operator-controller-manager-5f8c65bbfc-cptsj\" (UID: \"dc849752-f58a-4e7a-b1fa-5b7d88c14d1a\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.111940 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqthf\" (UniqueName: \"kubernetes.io/projected/ad9c89d4-9792-401b-9f1f-81d2e5c7ede9-kube-api-access-pqthf\") pod \"placement-operator-controller-manager-78f8948974-wqn2j\" (UID: \"ad9c89d4-9792-401b-9f1f-81d2e5c7ede9\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.115248 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.142266 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.143057 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.147063 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2kp5\" (UniqueName: \"kubernetes.io/projected/7965a5b4-89be-44c6-8471-20ede0f6c92f-kube-api-access-l2kp5\") pod \"test-operator-controller-manager-5854674fcc-nrrpt\" (UID: \"7965a5b4-89be-44c6-8471-20ede0f6c92f\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.161099 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.162034 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.165331 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.165707 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-v8h7l" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.165819 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.189478 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.209057 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.216231 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.218882 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.218985 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.227918 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8mwwr" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.249948 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.251138 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dw9m\" (UniqueName: \"kubernetes.io/projected/b71451df-d329-44b4-8d63-f0ea35db6ce7-kube-api-access-4dw9m\") pod \"watcher-operator-controller-manager-769dc69bc-rrnmn\" (UID: \"b71451df-d329-44b4-8d63-f0ea35db6ce7\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.251309 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2kp5\" (UniqueName: \"kubernetes.io/projected/7965a5b4-89be-44c6-8471-20ede0f6c92f-kube-api-access-l2kp5\") pod \"test-operator-controller-manager-5854674fcc-nrrpt\" (UID: \"7965a5b4-89be-44c6-8471-20ede0f6c92f\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.278824 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2kp5\" (UniqueName: \"kubernetes.io/projected/7965a5b4-89be-44c6-8471-20ede0f6c92f-kube-api-access-l2kp5\") pod \"test-operator-controller-manager-5854674fcc-nrrpt\" (UID: \"7965a5b4-89be-44c6-8471-20ede0f6c92f\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.363942 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swr5v\" (UniqueName: \"kubernetes.io/projected/6b03a06a-b4d5-4443-98eb-fdea3fd8302c-kube-api-access-swr5v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-ctlsq\" (UID: \"6b03a06a-b4d5-4443-98eb-fdea3fd8302c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.363989 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.364014 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-298lm\" (UniqueName: \"kubernetes.io/projected/e68bdc86-d09e-48ae-839b-9866e70e1383-kube-api-access-298lm\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.364034 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dw9m\" (UniqueName: \"kubernetes.io/projected/b71451df-d329-44b4-8d63-f0ea35db6ce7-kube-api-access-4dw9m\") pod \"watcher-operator-controller-manager-769dc69bc-rrnmn\" (UID: \"b71451df-d329-44b4-8d63-f0ea35db6ce7\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.364100 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.398709 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dw9m\" (UniqueName: \"kubernetes.io/projected/b71451df-d329-44b4-8d63-f0ea35db6ce7-kube-api-access-4dw9m\") pod \"watcher-operator-controller-manager-769dc69bc-rrnmn\" (UID: \"b71451df-d329-44b4-8d63-f0ea35db6ce7\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.407689 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.455394 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.467100 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.467190 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.467271 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swr5v\" (UniqueName: \"kubernetes.io/projected/6b03a06a-b4d5-4443-98eb-fdea3fd8302c-kube-api-access-swr5v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-ctlsq\" (UID: \"6b03a06a-b4d5-4443-98eb-fdea3fd8302c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.467307 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.467331 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-298lm\" (UniqueName: \"kubernetes.io/projected/e68bdc86-d09e-48ae-839b-9866e70e1383-kube-api-access-298lm\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.467652 4725 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.467713 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert podName:9c4065a8-50f2-4be9-bdd7-42345814f4c5 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:37.467696385 +0000 UTC m=+848.424338080 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" (UID: "9c4065a8-50f2-4be9-bdd7-42345814f4c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.468297 4725 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.468340 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:36.96832744 +0000 UTC m=+847.924969135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.468484 4725 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.468517 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:36.968507745 +0000 UTC m=+847.925149440 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "metrics-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.488135 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-298lm\" (UniqueName: \"kubernetes.io/projected/e68bdc86-d09e-48ae-839b-9866e70e1383-kube-api-access-298lm\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.493893 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swr5v\" (UniqueName: \"kubernetes.io/projected/6b03a06a-b4d5-4443-98eb-fdea3fd8302c-kube-api-access-swr5v\") pod \"rabbitmq-cluster-operator-manager-668c99d594-ctlsq\" (UID: \"6b03a06a-b4d5-4443-98eb-fdea3fd8302c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.529668 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.639917 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.656136 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.788834 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.798809 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" event={"ID":"c42118d3-322d-4733-8749-d13bc174a6f8","Type":"ContainerStarted","Data":"d2c4ebee0e0c71e8ee3a27d80ceae72ba5630e3103506dcf3df9466e21e10e70"} Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.801953 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" event={"ID":"65a9f532-387a-487e-bfe0-81a110509366","Type":"ContainerStarted","Data":"7cb43f30510a51cfb878fc8e70d319ad4770d72fc7342602e9841ed8ffe3e0a6"} Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.804279 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" event={"ID":"570a3a73-dffb-4ded-8fd1-d30ddc60979d","Type":"ContainerStarted","Data":"d45029cbd0111273c362e3d123e7ea588cba6260904b2729e7db65bc79bd3080"} Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.811518 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" event={"ID":"b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b","Type":"ContainerStarted","Data":"5c0d348e43f7ce6e4dd862a715523395f8f2de245105a7e6836a8747cbce60fb"} Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.814942 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nbhnj"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.816828 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.822515 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbhnj"] Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.872825 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqw9p\" (UniqueName: \"kubernetes.io/projected/f7a00dd9-d19f-415f-bad6-0114f989bdae-kube-api-access-qqw9p\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.873000 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7a00dd9-d19f-415f-bad6-0114f989bdae-catalog-content\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.873056 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7a00dd9-d19f-415f-bad6-0114f989bdae-utilities\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.974210 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.974592 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7a00dd9-d19f-415f-bad6-0114f989bdae-catalog-content\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.974635 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7a00dd9-d19f-415f-bad6-0114f989bdae-utilities\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.974710 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqw9p\" (UniqueName: \"kubernetes.io/projected/f7a00dd9-d19f-415f-bad6-0114f989bdae-kube-api-access-qqw9p\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.974754 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.974820 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.974400 4725 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.975055 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:37.975031812 +0000 UTC m=+848.931673507 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.975727 4725 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.975864 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:37.975843812 +0000 UTC m=+848.932485567 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "metrics-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.976054 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7a00dd9-d19f-415f-bad6-0114f989bdae-utilities\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: I1202 13:18:36.975763 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7a00dd9-d19f-415f-bad6-0114f989bdae-catalog-content\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.974986 4725 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:36 crc kubenswrapper[4725]: E1202 13:18:36.976120 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert podName:bf404326-3c43-4cc4-93b7-0793213afcba nodeName:}" failed. No retries permitted until 2025-12-02 13:18:38.976111719 +0000 UTC m=+849.932753414 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert") pod "infra-operator-controller-manager-57548d458d-g99db" (UID: "bf404326-3c43-4cc4-93b7-0793213afcba") : secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.002745 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqw9p\" (UniqueName: \"kubernetes.io/projected/f7a00dd9-d19f-415f-bad6-0114f989bdae-kube-api-access-qqw9p\") pod \"community-operators-nbhnj\" (UID: \"f7a00dd9-d19f-415f-bad6-0114f989bdae\") " pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.075702 4725 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.075760 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.109005 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.114429 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.119450 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6"] Dec 02 13:18:37 crc kubenswrapper[4725]: W1202 13:18:37.157587 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bad83be_e3ed_499a_8e64_cfe83c4ac1e6.slice/crio-11f009b859a1a6325b8833de4312e4990ea153dd16b576abd01e95c81a8502ca WatchSource:0}: Error finding container 11f009b859a1a6325b8833de4312e4990ea153dd16b576abd01e95c81a8502ca: Status 404 returned error can't find the container with id 11f009b859a1a6325b8833de4312e4990ea153dd16b576abd01e95c81a8502ca Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.157803 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.326706 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7w6j5" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.400964 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.417765 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-zdj64"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.417803 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q"] Dec 02 13:18:37 crc kubenswrapper[4725]: W1202 13:18:37.424653 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd2ee9ec_8397_4e64_8ea9_dc8f0466ab4b.slice/crio-edcf138312be1936ad2fe02886ba31f8cca386fbc72dfde938dbdb38ee3d3698 WatchSource:0}: Error finding container edcf138312be1936ad2fe02886ba31f8cca386fbc72dfde938dbdb38ee3d3698: Status 404 returned error can't find the container with id edcf138312be1936ad2fe02886ba31f8cca386fbc72dfde938dbdb38ee3d3698 Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.458165 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.475696 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.496815 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.497868 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.498144 4725 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.498203 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert podName:9c4065a8-50f2-4be9-bdd7-42345814f4c5 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:39.498190489 +0000 UTC m=+850.454832184 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" (UID: "9c4065a8-50f2-4be9-bdd7-42345814f4c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.503318 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.508657 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.516326 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.517101 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.526848 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj"] Dec 02 13:18:37 crc kubenswrapper[4725]: W1202 13:18:37.529172 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc849752_f58a_4e7a_b1fa_5b7d88c14d1a.slice/crio-2c60d57ed3074703bcd16e5a9621b535c7a52e42106e8e60f564817f67f7521b WatchSource:0}: Error finding container 2c60d57ed3074703bcd16e5a9621b535c7a52e42106e8e60f564817f67f7521b: Status 404 returned error can't find the container with id 2c60d57ed3074703bcd16e5a9621b535c7a52e42106e8e60f564817f67f7521b Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.530803 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq"] Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.532411 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mwjb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-cptsj_openstack-operators(dc849752-f58a-4e7a-b1fa-5b7d88c14d1a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.547340 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mwjb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-cptsj_openstack-operators(dc849752-f58a-4e7a-b1fa-5b7d88c14d1a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.547502 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7vwjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-p22fj_openstack-operators(3fb7e977-cd14-4b71-b349-19d4487cfb15): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: W1202 13:18:37.549224 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb71451df_d329_44b4_8d63_f0ea35db6ce7.slice/crio-f8049017dcece00a1712c20a8152d62e50bc0a4313435097454ad6ee18b10fd5 WatchSource:0}: Error finding container f8049017dcece00a1712c20a8152d62e50bc0a4313435097454ad6ee18b10fd5: Status 404 returned error can't find the container with id f8049017dcece00a1712c20a8152d62e50bc0a4313435097454ad6ee18b10fd5 Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.550554 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" podUID="dc849752-f58a-4e7a-b1fa-5b7d88c14d1a" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.565887 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4dw9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-rrnmn_openstack-operators(b71451df-d329-44b4-8d63-f0ea35db6ce7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.575104 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4dw9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-rrnmn_openstack-operators(b71451df-d329-44b4-8d63-f0ea35db6ce7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.576494 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" podUID="b71451df-d329-44b4-8d63-f0ea35db6ce7" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.576638 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swr5v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-ctlsq_openstack-operators(6b03a06a-b4d5-4443-98eb-fdea3fd8302c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.579530 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" podUID="6b03a06a-b4d5-4443-98eb-fdea3fd8302c" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.690491 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd"] Dec 02 13:18:37 crc kubenswrapper[4725]: W1202 13:18:37.700305 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod511bab3c_3885_4c5a_8c91_bd231b1c2a03.slice/crio-51dcda59053a22a0bde156b7caa4243e927ef03d0cdfad738753045005bbcf3b WatchSource:0}: Error finding container 51dcda59053a22a0bde156b7caa4243e927ef03d0cdfad738753045005bbcf3b: Status 404 returned error can't find the container with id 51dcda59053a22a0bde156b7caa4243e927ef03d0cdfad738753045005bbcf3b Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.703077 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvmcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mxpqd_openstack-operators(511bab3c-3885-4c5a-8c91-bd231b1c2a03): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.708558 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvmcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mxpqd_openstack-operators(511bab3c-3885-4c5a-8c91-bd231b1c2a03): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.709775 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podUID="511bab3c-3885-4c5a-8c91-bd231b1c2a03" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.825095 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" event={"ID":"4bad83be-e3ed-499a-8e64-cfe83c4ac1e6","Type":"ContainerStarted","Data":"11f009b859a1a6325b8833de4312e4990ea153dd16b576abd01e95c81a8502ca"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.827340 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" event={"ID":"bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b","Type":"ContainerStarted","Data":"edcf138312be1936ad2fe02886ba31f8cca386fbc72dfde938dbdb38ee3d3698"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.830559 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" event={"ID":"eacfb091-583b-4e89-9850-cc146cef7e20","Type":"ContainerStarted","Data":"2578251ef2e2e6ddf8dbbf58fbce49b07ab4832d51f99efb39f9915887d58ab2"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.835008 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" event={"ID":"6b08a8fe-08d5-4b04-9283-51973a9ac3d6","Type":"ContainerStarted","Data":"1962be9f899b7b0a5a689bd7cf8d7566ebe60509492ac89adb814f4a9f51f2e8"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.836945 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" event={"ID":"6b03a06a-b4d5-4443-98eb-fdea3fd8302c","Type":"ContainerStarted","Data":"948bc0257a7e8d200c65ae1ba7d0745cdb1e11a449cdfc03d9e924b679b8a288"} Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.838808 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" podUID="6b03a06a-b4d5-4443-98eb-fdea3fd8302c" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.839346 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" event={"ID":"dc849752-f58a-4e7a-b1fa-5b7d88c14d1a","Type":"ContainerStarted","Data":"2c60d57ed3074703bcd16e5a9621b535c7a52e42106e8e60f564817f67f7521b"} Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.844183 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" podUID="dc849752-f58a-4e7a-b1fa-5b7d88c14d1a" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.845104 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" event={"ID":"511bab3c-3885-4c5a-8c91-bd231b1c2a03","Type":"ContainerStarted","Data":"51dcda59053a22a0bde156b7caa4243e927ef03d0cdfad738753045005bbcf3b"} Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.847060 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podUID="511bab3c-3885-4c5a-8c91-bd231b1c2a03" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.847490 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" event={"ID":"3fb7e977-cd14-4b71-b349-19d4487cfb15","Type":"ContainerStarted","Data":"07b22ff96c3772ff699067deaee1cdbfcd37e76b37262d70033f7842a951053f"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.848298 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbhnj"] Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.854497 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" event={"ID":"8f32e184-607a-4077-aa41-ca7aabba76aa","Type":"ContainerStarted","Data":"cddf36e31ee72a83849f64c4095a8e5ea424802b73ee72ad9ce32090bd212060"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.856226 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" event={"ID":"bd74c0bd-9ef5-487c-b190-2b23a8120c8d","Type":"ContainerStarted","Data":"9dfb62dd53cd30453a9bfe2c8196ff28a2f875c871fa01a609a14277fd83c75e"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.858380 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" event={"ID":"ad9c89d4-9792-401b-9f1f-81d2e5c7ede9","Type":"ContainerStarted","Data":"3e9df08eb7801825f2762a871e246defadc8a3984b9c9d73bd16d2ab9b030d11"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.859758 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" event={"ID":"0b562f2a-162c-4611-81d3-00937e945f96","Type":"ContainerStarted","Data":"912e07c2801bc90db892b225e64d8e10bb009bda87ba6f2ae5addc487a2c4f9e"} Dec 02 13:18:37 crc kubenswrapper[4725]: W1202 13:18:37.863993 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7a00dd9_d19f_415f_bad6_0114f989bdae.slice/crio-87fcef9cf9233c32f24aaa4aae72b278f5afee2d78e73b3aca01fe8e9dc30f97 WatchSource:0}: Error finding container 87fcef9cf9233c32f24aaa4aae72b278f5afee2d78e73b3aca01fe8e9dc30f97: Status 404 returned error can't find the container with id 87fcef9cf9233c32f24aaa4aae72b278f5afee2d78e73b3aca01fe8e9dc30f97 Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.864647 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" event={"ID":"c1f336f6-716b-4096-89cd-f23d1fed90cb","Type":"ContainerStarted","Data":"7cb7bdbcb4ac3f131eaa2f5d9ae15809a87ca71d446af14381dc4bcb7c84abac"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.866343 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" event={"ID":"2c988ae2-9613-4584-a599-a7d063611189","Type":"ContainerStarted","Data":"dac2d1c3eef0108d877e1b74390fb409828a68f74a3bfc8d3b3e551ef690ee4c"} Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.867345 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" event={"ID":"b71451df-d329-44b4-8d63-f0ea35db6ce7","Type":"ContainerStarted","Data":"f8049017dcece00a1712c20a8152d62e50bc0a4313435097454ad6ee18b10fd5"} Dec 02 13:18:37 crc kubenswrapper[4725]: E1202 13:18:37.872567 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" podUID="b71451df-d329-44b4-8d63-f0ea35db6ce7" Dec 02 13:18:37 crc kubenswrapper[4725]: I1202 13:18:37.876225 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" event={"ID":"7965a5b4-89be-44c6-8471-20ede0f6c92f","Type":"ContainerStarted","Data":"d4fca43b059e510ddd604437b0f842285073f99c5f174309af69068d9e887db6"} Dec 02 13:18:38 crc kubenswrapper[4725]: I1202 13:18:38.009578 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:38 crc kubenswrapper[4725]: I1202 13:18:38.009683 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.009812 4725 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.009866 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:40.009849333 +0000 UTC m=+850.966491028 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "webhook-server-cert" not found Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.010043 4725 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.010141 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:40.01011524 +0000 UTC m=+850.966756965 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "metrics-server-cert" not found Dec 02 13:18:38 crc kubenswrapper[4725]: I1202 13:18:38.894034 4725 generic.go:334] "Generic (PLEG): container finished" podID="f7a00dd9-d19f-415f-bad6-0114f989bdae" containerID="a984363474897d8a4f75a1072ae8bf90c86c9f607d72ae0bda32e851462c3cd2" exitCode=0 Dec 02 13:18:38 crc kubenswrapper[4725]: I1202 13:18:38.895365 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbhnj" event={"ID":"f7a00dd9-d19f-415f-bad6-0114f989bdae","Type":"ContainerDied","Data":"a984363474897d8a4f75a1072ae8bf90c86c9f607d72ae0bda32e851462c3cd2"} Dec 02 13:18:38 crc kubenswrapper[4725]: I1202 13:18:38.895427 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbhnj" event={"ID":"f7a00dd9-d19f-415f-bad6-0114f989bdae","Type":"ContainerStarted","Data":"87fcef9cf9233c32f24aaa4aae72b278f5afee2d78e73b3aca01fe8e9dc30f97"} Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.898420 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podUID="511bab3c-3885-4c5a-8c91-bd231b1c2a03" Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.898558 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" podUID="6b03a06a-b4d5-4443-98eb-fdea3fd8302c" Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.899159 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" podUID="dc849752-f58a-4e7a-b1fa-5b7d88c14d1a" Dec 02 13:18:38 crc kubenswrapper[4725]: E1202 13:18:38.899262 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" podUID="b71451df-d329-44b4-8d63-f0ea35db6ce7" Dec 02 13:18:39 crc kubenswrapper[4725]: I1202 13:18:39.025701 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:39 crc kubenswrapper[4725]: E1202 13:18:39.025872 4725 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:39 crc kubenswrapper[4725]: E1202 13:18:39.025921 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert podName:bf404326-3c43-4cc4-93b7-0793213afcba nodeName:}" failed. No retries permitted until 2025-12-02 13:18:43.025905723 +0000 UTC m=+853.982547418 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert") pod "infra-operator-controller-manager-57548d458d-g99db" (UID: "bf404326-3c43-4cc4-93b7-0793213afcba") : secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:39 crc kubenswrapper[4725]: I1202 13:18:39.537561 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:39 crc kubenswrapper[4725]: E1202 13:18:39.537686 4725 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:39 crc kubenswrapper[4725]: E1202 13:18:39.537748 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert podName:9c4065a8-50f2-4be9-bdd7-42345814f4c5 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:43.537733241 +0000 UTC m=+854.494374936 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" (UID: "9c4065a8-50f2-4be9-bdd7-42345814f4c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:40 crc kubenswrapper[4725]: I1202 13:18:40.044172 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:40 crc kubenswrapper[4725]: I1202 13:18:40.044262 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:40 crc kubenswrapper[4725]: E1202 13:18:40.044381 4725 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 13:18:40 crc kubenswrapper[4725]: E1202 13:18:40.044396 4725 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 13:18:40 crc kubenswrapper[4725]: E1202 13:18:40.044440 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:44.044422873 +0000 UTC m=+855.001064568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "webhook-server-cert" not found Dec 02 13:18:40 crc kubenswrapper[4725]: E1202 13:18:40.044497 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:44.044474614 +0000 UTC m=+855.001116319 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "metrics-server-cert" not found Dec 02 13:18:43 crc kubenswrapper[4725]: I1202 13:18:43.101972 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:43 crc kubenswrapper[4725]: E1202 13:18:43.102199 4725 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:43 crc kubenswrapper[4725]: E1202 13:18:43.102522 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert podName:bf404326-3c43-4cc4-93b7-0793213afcba nodeName:}" failed. No retries permitted until 2025-12-02 13:18:51.102498054 +0000 UTC m=+862.059139749 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert") pod "infra-operator-controller-manager-57548d458d-g99db" (UID: "bf404326-3c43-4cc4-93b7-0793213afcba") : secret "infra-operator-webhook-server-cert" not found Dec 02 13:18:43 crc kubenswrapper[4725]: I1202 13:18:43.609616 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:43 crc kubenswrapper[4725]: E1202 13:18:43.609798 4725 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:43 crc kubenswrapper[4725]: E1202 13:18:43.609868 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert podName:9c4065a8-50f2-4be9-bdd7-42345814f4c5 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:51.609846872 +0000 UTC m=+862.566488567 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" (UID: "9c4065a8-50f2-4be9-bdd7-42345814f4c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 13:18:44 crc kubenswrapper[4725]: I1202 13:18:44.117742 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:44 crc kubenswrapper[4725]: I1202 13:18:44.117878 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:44 crc kubenswrapper[4725]: E1202 13:18:44.118186 4725 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 13:18:44 crc kubenswrapper[4725]: E1202 13:18:44.118246 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:52.118227445 +0000 UTC m=+863.074869150 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "metrics-server-cert" not found Dec 02 13:18:44 crc kubenswrapper[4725]: E1202 13:18:44.118298 4725 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 13:18:44 crc kubenswrapper[4725]: E1202 13:18:44.118322 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs podName:e68bdc86-d09e-48ae-839b-9866e70e1383 nodeName:}" failed. No retries permitted until 2025-12-02 13:18:52.118314547 +0000 UTC m=+863.074956242 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs") pod "openstack-operator-controller-manager-754459f956-vkwfn" (UID: "e68bdc86-d09e-48ae-839b-9866e70e1383") : secret "webhook-server-cert" not found Dec 02 13:18:50 crc kubenswrapper[4725]: E1202 13:18:50.831113 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 02 13:18:50 crc kubenswrapper[4725]: E1202 13:18:50.831951 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9dnx5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-ppgnw_openstack-operators(c42118d3-322d-4733-8749-d13bc174a6f8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:18:51 crc kubenswrapper[4725]: I1202 13:18:51.119907 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:51 crc kubenswrapper[4725]: I1202 13:18:51.126182 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf404326-3c43-4cc4-93b7-0793213afcba-cert\") pod \"infra-operator-controller-manager-57548d458d-g99db\" (UID: \"bf404326-3c43-4cc4-93b7-0793213afcba\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:51 crc kubenswrapper[4725]: I1202 13:18:51.410094 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:18:51 crc kubenswrapper[4725]: I1202 13:18:51.627953 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:51 crc kubenswrapper[4725]: I1202 13:18:51.631836 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9c4065a8-50f2-4be9-bdd7-42345814f4c5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc\" (UID: \"9c4065a8-50f2-4be9-bdd7-42345814f4c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:51 crc kubenswrapper[4725]: I1202 13:18:51.896066 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:18:52 crc kubenswrapper[4725]: I1202 13:18:52.135199 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:52 crc kubenswrapper[4725]: I1202 13:18:52.135291 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:52 crc kubenswrapper[4725]: I1202 13:18:52.139287 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-metrics-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:52 crc kubenswrapper[4725]: I1202 13:18:52.139846 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e68bdc86-d09e-48ae-839b-9866e70e1383-webhook-certs\") pod \"openstack-operator-controller-manager-754459f956-vkwfn\" (UID: \"e68bdc86-d09e-48ae-839b-9866e70e1383\") " pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:52 crc kubenswrapper[4725]: I1202 13:18:52.368041 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:18:53 crc kubenswrapper[4725]: E1202 13:18:53.485190 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 02 13:18:53 crc kubenswrapper[4725]: E1202 13:18:53.485576 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-szh7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-zdj64_openstack-operators(bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:18:55 crc kubenswrapper[4725]: E1202 13:18:55.968109 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 02 13:18:55 crc kubenswrapper[4725]: E1202 13:18:55.968652 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46tfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-wd49v_openstack-operators(2c988ae2-9613-4584-a599-a7d063611189): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:18:56 crc kubenswrapper[4725]: E1202 13:18:56.579499 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 02 13:18:56 crc kubenswrapper[4725]: E1202 13:18:56.579682 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t2xpk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-ctzjq_openstack-operators(b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:18:57 crc kubenswrapper[4725]: E1202 13:18:57.148778 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 02 13:18:57 crc kubenswrapper[4725]: E1202 13:18:57.148950 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fdzmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-rk99q_openstack-operators(c1f336f6-716b-4096-89cd-f23d1fed90cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:09 crc kubenswrapper[4725]: E1202 13:19:09.304014 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 02 13:19:09 crc kubenswrapper[4725]: E1202 13:19:09.304669 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ltc9k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-cllpp_openstack-operators(eacfb091-583b-4e89-9850-cc146cef7e20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:09 crc kubenswrapper[4725]: E1202 13:19:09.777390 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 02 13:19:09 crc kubenswrapper[4725]: E1202 13:19:09.777591 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4dw9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-rrnmn_openstack-operators(b71451df-d329-44b4-8d63-f0ea35db6ce7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:10 crc kubenswrapper[4725]: E1202 13:19:10.509973 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 02 13:19:10 crc kubenswrapper[4725]: E1202 13:19:10.510215 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvmcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mxpqd_openstack-operators(511bab3c-3885-4c5a-8c91-bd231b1c2a03): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:11 crc kubenswrapper[4725]: E1202 13:19:11.025924 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 02 13:19:11 crc kubenswrapper[4725]: E1202 13:19:11.026081 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mwjb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-cptsj_openstack-operators(dc849752-f58a-4e7a-b1fa-5b7d88c14d1a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:12 crc kubenswrapper[4725]: E1202 13:19:12.646055 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 02 13:19:12 crc kubenswrapper[4725]: E1202 13:19:12.646476 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-swr5v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-ctlsq_openstack-operators(6b03a06a-b4d5-4443-98eb-fdea3fd8302c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:12 crc kubenswrapper[4725]: E1202 13:19:12.647698 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" podUID="6b03a06a-b4d5-4443-98eb-fdea3fd8302c" Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.164403 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc"] Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.200255 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" event={"ID":"ad9c89d4-9792-401b-9f1f-81d2e5c7ede9","Type":"ContainerStarted","Data":"841a7079a1478255d5b80944e34b07fc45a6ec4f55feb98ce481356fd082cc72"} Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.203091 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" event={"ID":"0b562f2a-162c-4611-81d3-00937e945f96","Type":"ContainerStarted","Data":"c7fd16ab5b74e684713466de6adc56bfa653d8a30663a027553c780d2de0434a"} Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.212905 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" event={"ID":"8f32e184-607a-4077-aa41-ca7aabba76aa","Type":"ContainerStarted","Data":"2ca2ae895d6fa87f3e8e52661e506c1364ceeb38177ec9194eaffd6ff52c2bda"} Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.218070 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" event={"ID":"6b08a8fe-08d5-4b04-9283-51973a9ac3d6","Type":"ContainerStarted","Data":"cd0fb4de91e2e5f2586a8489dba8df4c191751400b06eaa100d90693b097199f"} Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.226527 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" event={"ID":"4bad83be-e3ed-499a-8e64-cfe83c4ac1e6","Type":"ContainerStarted","Data":"b6c532789d828700e3be41945af511755fa1e65d34e7dd050f049b5e96049128"} Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.231163 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.231308 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9dnx5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-ppgnw_openstack-operators(c42118d3-322d-4733-8749-d13bc174a6f8): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.232031 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" event={"ID":"65a9f532-387a-487e-bfe0-81a110509366","Type":"ContainerStarted","Data":"aca918e074f6f62a2a376cee043a3b4575f3b64d7c24b57e7393ffb7f878b6ac"} Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.235858 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" podUID="c42118d3-322d-4733-8749-d13bc174a6f8" Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.236845 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" event={"ID":"7965a5b4-89be-44c6-8471-20ede0f6c92f","Type":"ContainerStarted","Data":"22301d52b880c93cda98422dcaae3092e3ed60f369571bb3109f2f1f67aaa05b"} Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.241877 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g99db"] Dec 02 13:19:13 crc kubenswrapper[4725]: I1202 13:19:13.265404 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn"] Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.266784 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.266940 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t2xpk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-ctzjq_openstack-operators(b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.268254 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" podUID="b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.408592 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.409064 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46tfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-wd49v_openstack-operators(2c988ae2-9613-4584-a599-a7d063611189): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.410270 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" podUID="2c988ae2-9613-4584-a599-a7d063611189" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.430403 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.430614 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-szh7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-zdj64_openstack-operators(bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 02 13:19:13 crc kubenswrapper[4725]: E1202 13:19:13.431772 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" podUID="bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b" Dec 02 13:19:14 crc kubenswrapper[4725]: I1202 13:19:14.249607 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" event={"ID":"570a3a73-dffb-4ded-8fd1-d30ddc60979d","Type":"ContainerStarted","Data":"08b174483b2ea821bbe999709ed4fc086580c95ffc3fc8fd52bb0446238fdd8b"} Dec 02 13:19:14 crc kubenswrapper[4725]: I1202 13:19:14.250876 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" event={"ID":"bf404326-3c43-4cc4-93b7-0793213afcba","Type":"ContainerStarted","Data":"883ce64fbd8855d37727c2892c06c138f7f7a52493d8d7a391bd26c825bd1781"} Dec 02 13:19:14 crc kubenswrapper[4725]: I1202 13:19:14.252926 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" event={"ID":"e68bdc86-d09e-48ae-839b-9866e70e1383","Type":"ContainerStarted","Data":"0f4ea19ddb2eb7b319ea3cc1c8018305e58e7a4dc6ec0c3f513f27c2cccc3cec"} Dec 02 13:19:14 crc kubenswrapper[4725]: I1202 13:19:14.254218 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" event={"ID":"9c4065a8-50f2-4be9-bdd7-42345814f4c5","Type":"ContainerStarted","Data":"5a0505f01cb0094bd8ae77219fb1ea06c4a31361b6826e6c6b21e7722dac0ca0"} Dec 02 13:19:14 crc kubenswrapper[4725]: I1202 13:19:14.256130 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" event={"ID":"bd74c0bd-9ef5-487c-b190-2b23a8120c8d","Type":"ContainerStarted","Data":"e1b754e967e0d451bf60dbbd9f787bbc0dac0ed308addd8ca46623dcb6812e74"} Dec 02 13:19:15 crc kubenswrapper[4725]: E1202 13:19:15.369955 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 02 13:19:15 crc kubenswrapper[4725]: E1202 13:19:15.370192 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7vwjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-p22fj_openstack-operators(3fb7e977-cd14-4b71-b349-19d4487cfb15): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:19:15 crc kubenswrapper[4725]: E1202 13:19:15.371427 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" podUID="3fb7e977-cd14-4b71-b349-19d4487cfb15" Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.284327 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" event={"ID":"7965a5b4-89be-44c6-8471-20ede0f6c92f","Type":"ContainerStarted","Data":"bdea8422e462196cd3915c05f349758fe26a104473ce7e66751530ebe8c541f6"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.285564 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.286636 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" event={"ID":"c42118d3-322d-4733-8749-d13bc174a6f8","Type":"ContainerStarted","Data":"280b559f795ff9ee0ccb0ce98bfe19197f76dc66961ec699824bddb96c574c99"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.287984 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" event={"ID":"2c988ae2-9613-4584-a599-a7d063611189","Type":"ContainerStarted","Data":"07516299d1a8d16d1b22a72d549928523c9f404f94d8e99af0610ce1fc138576"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.304600 4725 generic.go:334] "Generic (PLEG): container finished" podID="f7a00dd9-d19f-415f-bad6-0114f989bdae" containerID="364593fbdb71b01ea0fdda30ca90b3018dcd7cf9ec4db369e988ba4c7b9b0fd0" exitCode=0 Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.304677 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbhnj" event={"ID":"f7a00dd9-d19f-415f-bad6-0114f989bdae","Type":"ContainerDied","Data":"364593fbdb71b01ea0fdda30ca90b3018dcd7cf9ec4db369e988ba4c7b9b0fd0"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.320836 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" event={"ID":"4bad83be-e3ed-499a-8e64-cfe83c4ac1e6","Type":"ContainerStarted","Data":"4414ea53ba03aae590c83b2a519827777810ed6b2ff8b3a603b1d8c79413a3ba"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.321093 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.327680 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" event={"ID":"b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b","Type":"ContainerStarted","Data":"f770bf2ed5c6c506d2620f23f940a39bfe47ac616a9c9122b2137fee36cab058"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.331043 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" event={"ID":"e68bdc86-d09e-48ae-839b-9866e70e1383","Type":"ContainerStarted","Data":"02f4f84172d14d394f84a3c7f97d3df1feeb0119dbcfa04d4fc4df739837ee97"} Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.331164 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.341381 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" podStartSLOduration=3.171454541 podStartE2EDuration="41.341360679s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.526494538 +0000 UTC m=+848.483136233" lastFinishedPulling="2025-12-02 13:19:15.696400676 +0000 UTC m=+886.653042371" observedRunningTime="2025-12-02 13:19:16.309597205 +0000 UTC m=+887.266238900" watchObservedRunningTime="2025-12-02 13:19:16.341360679 +0000 UTC m=+887.298002374" Dec 02 13:19:16 crc kubenswrapper[4725]: E1202 13:19:16.362656 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" podUID="eacfb091-583b-4e89-9850-cc146cef7e20" Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.369609 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" podStartSLOduration=40.369592266 podStartE2EDuration="40.369592266s" podCreationTimestamp="2025-12-02 13:18:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:19:16.367430982 +0000 UTC m=+887.324072677" watchObservedRunningTime="2025-12-02 13:19:16.369592266 +0000 UTC m=+887.326233961" Dec 02 13:19:16 crc kubenswrapper[4725]: E1202 13:19:16.440294 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" podUID="b71451df-d329-44b4-8d63-f0ea35db6ce7" Dec 02 13:19:16 crc kubenswrapper[4725]: I1202 13:19:16.466605 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" podStartSLOduration=2.920856939 podStartE2EDuration="41.466585839s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.1950074 +0000 UTC m=+848.151649095" lastFinishedPulling="2025-12-02 13:19:15.7407363 +0000 UTC m=+886.697377995" observedRunningTime="2025-12-02 13:19:16.461997866 +0000 UTC m=+887.418639561" watchObservedRunningTime="2025-12-02 13:19:16.466585839 +0000 UTC m=+887.423227534" Dec 02 13:19:16 crc kubenswrapper[4725]: E1202 13:19:16.625905 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" podUID="dc849752-f58a-4e7a-b1fa-5b7d88c14d1a" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.490907 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" event={"ID":"6b08a8fe-08d5-4b04-9283-51973a9ac3d6","Type":"ContainerStarted","Data":"181612dec43759b339cdadf32f40aad4e48ede54c91de7568a9ce38df26534e5"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.491169 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.498848 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" event={"ID":"65a9f532-387a-487e-bfe0-81a110509366","Type":"ContainerStarted","Data":"a7e6a04b0f2e98de4311e30a156f565662b06c4700824b5180e877e941de3500"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.498981 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.516310 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" podStartSLOduration=3.923509816 podStartE2EDuration="42.516297447s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.157639108 +0000 UTC m=+848.114280803" lastFinishedPulling="2025-12-02 13:19:15.750426749 +0000 UTC m=+886.707068434" observedRunningTime="2025-12-02 13:19:17.514170195 +0000 UTC m=+888.470811890" watchObservedRunningTime="2025-12-02 13:19:17.516297447 +0000 UTC m=+888.472939132" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.516721 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" event={"ID":"0b562f2a-162c-4611-81d3-00937e945f96","Type":"ContainerStarted","Data":"66b31a9b88c87cafeefa0326d2a53ac98599e645efcedce1720c142355fd3875"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.517427 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.528979 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.531886 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" event={"ID":"ad9c89d4-9792-401b-9f1f-81d2e5c7ede9","Type":"ContainerStarted","Data":"f78e3f13170d0f1516fb3fa52cd4c6c57cd5a41d100d5e4ebfc535e56767fff2"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.532581 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.534265 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" event={"ID":"8f32e184-607a-4077-aa41-ca7aabba76aa","Type":"ContainerStarted","Data":"23a8af770e8bec7fc2b0df236e53c9a7216ab4772a46891a39e4283a96b06959"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.538446 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.541180 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.544171 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" event={"ID":"eacfb091-583b-4e89-9850-cc146cef7e20","Type":"ContainerStarted","Data":"3e4180b6c32e1f2a92e3ae03d3796134636f23b0b2cfed9517c1334280106bf1"} Dec 02 13:19:17 crc kubenswrapper[4725]: E1202 13:19:17.545857 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" podUID="eacfb091-583b-4e89-9850-cc146cef7e20" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.546182 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" event={"ID":"b71451df-d329-44b4-8d63-f0ea35db6ce7","Type":"ContainerStarted","Data":"b64167c2c929fb1347414d6344be470a2c80514304f76c1e0afe494b807bf1b5"} Dec 02 13:19:17 crc kubenswrapper[4725]: E1202 13:19:17.547914 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" podUID="b71451df-d329-44b4-8d63-f0ea35db6ce7" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.553839 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" podStartSLOduration=4.513340966 podStartE2EDuration="43.553823243s" podCreationTimestamp="2025-12-02 13:18:34 +0000 UTC" firstStartedPulling="2025-12-02 13:18:36.7009583 +0000 UTC m=+847.657599995" lastFinishedPulling="2025-12-02 13:19:15.741440577 +0000 UTC m=+886.698082272" observedRunningTime="2025-12-02 13:19:17.550836189 +0000 UTC m=+888.507477904" watchObservedRunningTime="2025-12-02 13:19:17.553823243 +0000 UTC m=+888.510464938" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.557653 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" event={"ID":"dc849752-f58a-4e7a-b1fa-5b7d88c14d1a","Type":"ContainerStarted","Data":"aa9078e7dabe906ad62badc3f9b30542fe272c9c43d92134815323113f4f81c7"} Dec 02 13:19:17 crc kubenswrapper[4725]: E1202 13:19:17.558985 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" podUID="dc849752-f58a-4e7a-b1fa-5b7d88c14d1a" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.563343 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" event={"ID":"570a3a73-dffb-4ded-8fd1-d30ddc60979d","Type":"ContainerStarted","Data":"8234eed4e6fde80afbddb410144b377f68b0a9f6ab75597aad81a5ef5b1fbb76"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.564005 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.571889 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" event={"ID":"bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b","Type":"ContainerStarted","Data":"558dce30762d3c9437735e8ce988d01d154bc089147f800debbde0b9a595d8ba"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.585605 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" event={"ID":"bd74c0bd-9ef5-487c-b190-2b23a8120c8d","Type":"ContainerStarted","Data":"0e469b8c5482989ba09e20bef80f906d914a4696115e0109b9094d136afc84f7"} Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.585653 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.607477 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-8s44s" podStartSLOduration=4.255922726 podStartE2EDuration="42.607441616s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.457500455 +0000 UTC m=+848.414142150" lastFinishedPulling="2025-12-02 13:19:15.809019345 +0000 UTC m=+886.765661040" observedRunningTime="2025-12-02 13:19:17.576075572 +0000 UTC m=+888.532717267" watchObservedRunningTime="2025-12-02 13:19:17.607441616 +0000 UTC m=+888.564083311" Dec 02 13:19:17 crc kubenswrapper[4725]: E1202 13:19:17.625759 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" podUID="c1f336f6-716b-4096-89cd-f23d1fed90cb" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.650847 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-cwfk4" podStartSLOduration=4.09589237 podStartE2EDuration="42.650823867s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.194928198 +0000 UTC m=+848.151569893" lastFinishedPulling="2025-12-02 13:19:15.749859695 +0000 UTC m=+886.706501390" observedRunningTime="2025-12-02 13:19:17.648640373 +0000 UTC m=+888.605282068" watchObservedRunningTime="2025-12-02 13:19:17.650823867 +0000 UTC m=+888.607465562" Dec 02 13:19:17 crc kubenswrapper[4725]: E1202 13:19:17.663292 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podUID="511bab3c-3885-4c5a-8c91-bd231b1c2a03" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.743641 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" podStartSLOduration=4.521881059 podStartE2EDuration="42.743622736s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.531600304 +0000 UTC m=+848.488241999" lastFinishedPulling="2025-12-02 13:19:15.753341981 +0000 UTC m=+886.709983676" observedRunningTime="2025-12-02 13:19:17.722271699 +0000 UTC m=+888.678913394" watchObservedRunningTime="2025-12-02 13:19:17.743622736 +0000 UTC m=+888.700264431" Dec 02 13:19:17 crc kubenswrapper[4725]: I1202 13:19:17.837424 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" podStartSLOduration=4.59414805 podStartE2EDuration="43.83739805s" podCreationTimestamp="2025-12-02 13:18:34 +0000 UTC" firstStartedPulling="2025-12-02 13:18:36.552046696 +0000 UTC m=+847.508688381" lastFinishedPulling="2025-12-02 13:19:15.795296686 +0000 UTC m=+886.751938381" observedRunningTime="2025-12-02 13:19:17.836270962 +0000 UTC m=+888.792912657" watchObservedRunningTime="2025-12-02 13:19:17.83739805 +0000 UTC m=+888.794039745" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:17.920696 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" podStartSLOduration=4.699993353 podStartE2EDuration="42.920676135s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.529166293 +0000 UTC m=+848.485807988" lastFinishedPulling="2025-12-02 13:19:15.749849075 +0000 UTC m=+886.706490770" observedRunningTime="2025-12-02 13:19:17.914860061 +0000 UTC m=+888.871501756" watchObservedRunningTime="2025-12-02 13:19:17.920676135 +0000 UTC m=+888.877317830" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.732537 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" event={"ID":"2c988ae2-9613-4584-a599-a7d063611189","Type":"ContainerStarted","Data":"50eb6e7ea5472f091be28c069abe5deb03b0f90e7ea4696dc8d8af37b1390d00"} Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.733741 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.734715 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbhnj" event={"ID":"f7a00dd9-d19f-415f-bad6-0114f989bdae","Type":"ContainerStarted","Data":"5c2a81f51e1f30c22742361eb7ab32fa55cf53fd539bdae14ddab8ed66e9d3bb"} Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.746262 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" event={"ID":"b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b","Type":"ContainerStarted","Data":"187f1218c591632d28e5bf19a7895ffaf309b7fa973b34cbdac5642dd8767cef"} Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.746641 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.752824 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" event={"ID":"bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b","Type":"ContainerStarted","Data":"3f63e7131cdf6508a1eb6e8b0fc047a02375246709c42d23daf259774ce053be"} Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.753105 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.757890 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" podStartSLOduration=5.672820626 podStartE2EDuration="43.757876261s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.421587799 +0000 UTC m=+848.378229494" lastFinishedPulling="2025-12-02 13:19:15.506643434 +0000 UTC m=+886.463285129" observedRunningTime="2025-12-02 13:19:18.754478447 +0000 UTC m=+889.711120142" watchObservedRunningTime="2025-12-02 13:19:18.757876261 +0000 UTC m=+889.714517956" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.764762 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" event={"ID":"c42118d3-322d-4733-8749-d13bc174a6f8","Type":"ContainerStarted","Data":"d8ad0e25a63fe9fd8ad23b0b2c2c9a5f9988199093fae6afd0935e426cf6ed7c"} Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.764936 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.767188 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" event={"ID":"511bab3c-3885-4c5a-8c91-bd231b1c2a03","Type":"ContainerStarted","Data":"3a1f6177e653db4a3d5a16fc2c3471440e38c8df117d6d32d1f0ffc189aa566d"} Dec 02 13:19:18 crc kubenswrapper[4725]: E1202 13:19:18.769280 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podUID="511bab3c-3885-4c5a-8c91-bd231b1c2a03" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.769494 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" event={"ID":"c1f336f6-716b-4096-89cd-f23d1fed90cb","Type":"ContainerStarted","Data":"8a141c58a1a94a8e4c9d488bdef0c2ef96cb1a5794b80552e995b4b3f5aa947d"} Dec 02 13:19:18 crc kubenswrapper[4725]: E1202 13:19:18.771004 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" podUID="eacfb091-583b-4e89-9850-cc146cef7e20" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.771016 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" podStartSLOduration=5.317501227 podStartE2EDuration="44.771001575s" podCreationTimestamp="2025-12-02 13:18:34 +0000 UTC" firstStartedPulling="2025-12-02 13:18:36.175506205 +0000 UTC m=+847.132147900" lastFinishedPulling="2025-12-02 13:19:15.629006553 +0000 UTC m=+886.585648248" observedRunningTime="2025-12-02 13:19:18.770251996 +0000 UTC m=+889.726893691" watchObservedRunningTime="2025-12-02 13:19:18.771001575 +0000 UTC m=+889.727643270" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.774695 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nrrpt" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.774765 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-njlj7" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.774916 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-pmdw6" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.775370 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-2l2x4" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.775430 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wqn2j" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.778801 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-z9ttg" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.802519 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" podStartSLOduration=5.485007732 podStartE2EDuration="43.802503102s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.439602914 +0000 UTC m=+848.396244859" lastFinishedPulling="2025-12-02 13:19:15.757098534 +0000 UTC m=+886.713740229" observedRunningTime="2025-12-02 13:19:18.796886573 +0000 UTC m=+889.753528268" watchObservedRunningTime="2025-12-02 13:19:18.802503102 +0000 UTC m=+889.759144797" Dec 02 13:19:18 crc kubenswrapper[4725]: I1202 13:19:18.822401 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nbhnj" podStartSLOduration=4.533251474 podStartE2EDuration="42.822371502s" podCreationTimestamp="2025-12-02 13:18:36 +0000 UTC" firstStartedPulling="2025-12-02 13:18:38.898197782 +0000 UTC m=+849.854839477" lastFinishedPulling="2025-12-02 13:19:17.18731781 +0000 UTC m=+888.143959505" observedRunningTime="2025-12-02 13:19:18.819809299 +0000 UTC m=+889.776450994" watchObservedRunningTime="2025-12-02 13:19:18.822371502 +0000 UTC m=+889.779013197" Dec 02 13:19:19 crc kubenswrapper[4725]: I1202 13:19:19.157418 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" podStartSLOduration=6.319785056 podStartE2EDuration="45.157399628s" podCreationTimestamp="2025-12-02 13:18:34 +0000 UTC" firstStartedPulling="2025-12-02 13:18:36.698329505 +0000 UTC m=+847.654971200" lastFinishedPulling="2025-12-02 13:19:15.535944077 +0000 UTC m=+886.492585772" observedRunningTime="2025-12-02 13:19:19.152084867 +0000 UTC m=+890.108726562" watchObservedRunningTime="2025-12-02 13:19:19.157399628 +0000 UTC m=+890.114041333" Dec 02 13:19:20 crc kubenswrapper[4725]: I1202 13:19:20.825752 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" event={"ID":"c1f336f6-716b-4096-89cd-f23d1fed90cb","Type":"ContainerStarted","Data":"862f3a92e569787dd955d4abd1cefe0f2eb8175286a96c791b2537010c947596"} Dec 02 13:19:20 crc kubenswrapper[4725]: I1202 13:19:20.845325 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-wd49v" Dec 02 13:19:20 crc kubenswrapper[4725]: I1202 13:19:20.849520 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" podStartSLOduration=3.98852568 podStartE2EDuration="45.849504897s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.457166697 +0000 UTC m=+848.413808382" lastFinishedPulling="2025-12-02 13:19:19.318145904 +0000 UTC m=+890.274787599" observedRunningTime="2025-12-02 13:19:20.845347784 +0000 UTC m=+891.801989479" watchObservedRunningTime="2025-12-02 13:19:20.849504897 +0000 UTC m=+891.806146592" Dec 02 13:19:21 crc kubenswrapper[4725]: I1202 13:19:21.833897 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.283307 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mg728"] Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.287552 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.291514 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mg728"] Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.364721 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-utilities\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.364787 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqhxx\" (UniqueName: \"kubernetes.io/projected/b7b2fb09-935a-4165-ac85-82806d1a9f10-kube-api-access-hqhxx\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.364858 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-catalog-content\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.375074 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-754459f956-vkwfn" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.466149 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-catalog-content\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.466261 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-utilities\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.466340 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqhxx\" (UniqueName: \"kubernetes.io/projected/b7b2fb09-935a-4165-ac85-82806d1a9f10-kube-api-access-hqhxx\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.466802 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-catalog-content\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.466859 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-utilities\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.487821 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqhxx\" (UniqueName: \"kubernetes.io/projected/b7b2fb09-935a-4165-ac85-82806d1a9f10-kube-api-access-hqhxx\") pod \"certified-operators-mg728\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:22 crc kubenswrapper[4725]: I1202 13:19:22.619366 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:23 crc kubenswrapper[4725]: E1202 13:19:23.273566 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" podUID="6b03a06a-b4d5-4443-98eb-fdea3fd8302c" Dec 02 13:19:24 crc kubenswrapper[4725]: I1202 13:19:24.284746 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:19:24 crc kubenswrapper[4725]: I1202 13:19:24.284849 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:19:25 crc kubenswrapper[4725]: I1202 13:19:25.360275 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-ctzjq" Dec 02 13:19:25 crc kubenswrapper[4725]: I1202 13:19:25.360323 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-ppgnw" Dec 02 13:19:25 crc kubenswrapper[4725]: I1202 13:19:25.658073 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-klrxl" Dec 02 13:19:25 crc kubenswrapper[4725]: I1202 13:19:25.950054 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-zdj64" Dec 02 13:19:26 crc kubenswrapper[4725]: I1202 13:19:26.034689 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-rk99q" Dec 02 13:19:27 crc kubenswrapper[4725]: I1202 13:19:27.158742 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:19:27 crc kubenswrapper[4725]: I1202 13:19:27.159067 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:19:27 crc kubenswrapper[4725]: I1202 13:19:27.222821 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:19:28 crc kubenswrapper[4725]: I1202 13:19:28.024662 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" event={"ID":"bf404326-3c43-4cc4-93b7-0793213afcba","Type":"ContainerStarted","Data":"8fce447e7b79085103f1b8c3b810120366ec4fb41822bd57a206d3a12c632c25"} Dec 02 13:19:28 crc kubenswrapper[4725]: I1202 13:19:28.116476 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nbhnj" Dec 02 13:19:28 crc kubenswrapper[4725]: I1202 13:19:28.202962 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbhnj"] Dec 02 13:19:28 crc kubenswrapper[4725]: I1202 13:19:28.325738 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sg828"] Dec 02 13:19:28 crc kubenswrapper[4725]: I1202 13:19:28.326056 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sg828" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="registry-server" containerID="cri-o://45711932e9ff315ab326cc4df554f6f4aaac3b7181b337652230df4722f777f7" gracePeriod=2 Dec 02 13:19:28 crc kubenswrapper[4725]: I1202 13:19:28.373073 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mg728"] Dec 02 13:19:28 crc kubenswrapper[4725]: W1202 13:19:28.382381 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7b2fb09_935a_4165_ac85_82806d1a9f10.slice/crio-2e36ce2b0702996027ca69d66b2cef02d35b98fbefeba6d65a432220f7a59509 WatchSource:0}: Error finding container 2e36ce2b0702996027ca69d66b2cef02d35b98fbefeba6d65a432220f7a59509: Status 404 returned error can't find the container with id 2e36ce2b0702996027ca69d66b2cef02d35b98fbefeba6d65a432220f7a59509 Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.034628 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" event={"ID":"bf404326-3c43-4cc4-93b7-0793213afcba","Type":"ContainerStarted","Data":"51ed1fffdbf59a221c397260d63e4a4b914a7fdd1ab8a613bb5dc9f88518c2f4"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.034885 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.038339 4725 generic.go:334] "Generic (PLEG): container finished" podID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerID="6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7" exitCode=0 Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.038393 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mg728" event={"ID":"b7b2fb09-935a-4165-ac85-82806d1a9f10","Type":"ContainerDied","Data":"6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.038484 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mg728" event={"ID":"b7b2fb09-935a-4165-ac85-82806d1a9f10","Type":"ContainerStarted","Data":"2e36ce2b0702996027ca69d66b2cef02d35b98fbefeba6d65a432220f7a59509"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.041444 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" event={"ID":"9c4065a8-50f2-4be9-bdd7-42345814f4c5","Type":"ContainerStarted","Data":"4f2972ae7e8540ce5a1c5e02ab81ef3ec456ad9e82daccd671ceb654867d99e6"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.041509 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" event={"ID":"9c4065a8-50f2-4be9-bdd7-42345814f4c5","Type":"ContainerStarted","Data":"a8f89a7576731b3c97f78e980c4b8382f42329c1e6cc292d17c80150ec7771aa"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.042074 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.049103 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" event={"ID":"3fb7e977-cd14-4b71-b349-19d4487cfb15","Type":"ContainerStarted","Data":"b176739efb23152bab38256384f6854253e0438eef6cbc669a5f46862ed4d28f"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.049175 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" event={"ID":"3fb7e977-cd14-4b71-b349-19d4487cfb15","Type":"ContainerStarted","Data":"b98db3acf5e43850d915c1ad311fb2a399d675a6f49e5a5e45a9484c182cb722"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.049268 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.052057 4725 generic.go:334] "Generic (PLEG): container finished" podID="d64eb153-6898-4b22-b86b-df77e2c71044" containerID="45711932e9ff315ab326cc4df554f6f4aaac3b7181b337652230df4722f777f7" exitCode=0 Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.052546 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg828" event={"ID":"d64eb153-6898-4b22-b86b-df77e2c71044","Type":"ContainerDied","Data":"45711932e9ff315ab326cc4df554f6f4aaac3b7181b337652230df4722f777f7"} Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.060482 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" podStartSLOduration=39.600298522 podStartE2EDuration="54.060451854s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:19:13.269430925 +0000 UTC m=+884.226072620" lastFinishedPulling="2025-12-02 13:19:27.729584257 +0000 UTC m=+898.686225952" observedRunningTime="2025-12-02 13:19:29.056600489 +0000 UTC m=+900.013242184" watchObservedRunningTime="2025-12-02 13:19:29.060451854 +0000 UTC m=+900.017093559" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.117226 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" podStartSLOduration=39.649160567 podStartE2EDuration="54.117206024s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:19:13.261673564 +0000 UTC m=+884.218315269" lastFinishedPulling="2025-12-02 13:19:27.729719011 +0000 UTC m=+898.686360726" observedRunningTime="2025-12-02 13:19:29.115797519 +0000 UTC m=+900.072439234" watchObservedRunningTime="2025-12-02 13:19:29.117206024 +0000 UTC m=+900.073847719" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.140012 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" podStartSLOduration=3.946515223 podStartE2EDuration="54.139991646s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.547397513 +0000 UTC m=+848.504039208" lastFinishedPulling="2025-12-02 13:19:27.740873936 +0000 UTC m=+898.697515631" observedRunningTime="2025-12-02 13:19:29.138614692 +0000 UTC m=+900.095256387" watchObservedRunningTime="2025-12-02 13:19:29.139991646 +0000 UTC m=+900.096633441" Dec 02 13:19:29 crc kubenswrapper[4725]: E1202 13:19:29.276889 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podUID="511bab3c-3885-4c5a-8c91-bd231b1c2a03" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.277322 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.401805 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-utilities\") pod \"d64eb153-6898-4b22-b86b-df77e2c71044\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.401879 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvwkt\" (UniqueName: \"kubernetes.io/projected/d64eb153-6898-4b22-b86b-df77e2c71044-kube-api-access-mvwkt\") pod \"d64eb153-6898-4b22-b86b-df77e2c71044\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.401977 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-catalog-content\") pod \"d64eb153-6898-4b22-b86b-df77e2c71044\" (UID: \"d64eb153-6898-4b22-b86b-df77e2c71044\") " Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.403852 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-utilities" (OuterVolumeSpecName: "utilities") pod "d64eb153-6898-4b22-b86b-df77e2c71044" (UID: "d64eb153-6898-4b22-b86b-df77e2c71044"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.412775 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d64eb153-6898-4b22-b86b-df77e2c71044-kube-api-access-mvwkt" (OuterVolumeSpecName: "kube-api-access-mvwkt") pod "d64eb153-6898-4b22-b86b-df77e2c71044" (UID: "d64eb153-6898-4b22-b86b-df77e2c71044"). InnerVolumeSpecName "kube-api-access-mvwkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.478336 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d64eb153-6898-4b22-b86b-df77e2c71044" (UID: "d64eb153-6898-4b22-b86b-df77e2c71044"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.543667 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvwkt\" (UniqueName: \"kubernetes.io/projected/d64eb153-6898-4b22-b86b-df77e2c71044-kube-api-access-mvwkt\") on node \"crc\" DevicePath \"\"" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.543710 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.543723 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d64eb153-6898-4b22-b86b-df77e2c71044-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.642915 4725 scope.go:117] "RemoveContainer" containerID="f1041bd51ae6778c41539a821ad5e17054386be65b2136ddc793d2956a797a9f" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.731744 4725 scope.go:117] "RemoveContainer" containerID="372daf2d90a65c03b4b0661fb65af4ba15d86e063c397b9921093d2e144df547" Dec 02 13:19:29 crc kubenswrapper[4725]: I1202 13:19:29.878104 4725 scope.go:117] "RemoveContainer" containerID="45711932e9ff315ab326cc4df554f6f4aaac3b7181b337652230df4722f777f7" Dec 02 13:19:30 crc kubenswrapper[4725]: I1202 13:19:30.058563 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sg828" event={"ID":"d64eb153-6898-4b22-b86b-df77e2c71044","Type":"ContainerDied","Data":"2aa0aaa86c9a44c555e96f8aa374e678a9d67e5045217d734386ff22f12e0228"} Dec 02 13:19:30 crc kubenswrapper[4725]: I1202 13:19:30.058720 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sg828" Dec 02 13:19:30 crc kubenswrapper[4725]: I1202 13:19:30.100925 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sg828"] Dec 02 13:19:30 crc kubenswrapper[4725]: I1202 13:19:30.108192 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sg828"] Dec 02 13:19:31 crc kubenswrapper[4725]: I1202 13:19:31.067347 4725 generic.go:334] "Generic (PLEG): container finished" podID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerID="f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a" exitCode=0 Dec 02 13:19:31 crc kubenswrapper[4725]: I1202 13:19:31.067399 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mg728" event={"ID":"b7b2fb09-935a-4165-ac85-82806d1a9f10","Type":"ContainerDied","Data":"f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a"} Dec 02 13:19:31 crc kubenswrapper[4725]: I1202 13:19:31.276757 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" path="/var/lib/kubelet/pods/d64eb153-6898-4b22-b86b-df77e2c71044/volumes" Dec 02 13:19:32 crc kubenswrapper[4725]: I1202 13:19:32.270054 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:19:36 crc kubenswrapper[4725]: I1202 13:19:36.052186 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p22fj" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.129108 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" event={"ID":"b71451df-d329-44b4-8d63-f0ea35db6ce7","Type":"ContainerStarted","Data":"852adc9a69cdbe2a8684f137fd5ef291b77a86c9a6d550299a06aaaf15879909"} Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.130824 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" event={"ID":"dc849752-f58a-4e7a-b1fa-5b7d88c14d1a","Type":"ContainerStarted","Data":"e63fbbc395dbbe999b444c3b98672c3513e03dcc1c747214142c44fb7e50fa39"} Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.131215 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.133096 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mg728" event={"ID":"b7b2fb09-935a-4165-ac85-82806d1a9f10","Type":"ContainerStarted","Data":"f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af"} Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.134853 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" event={"ID":"eacfb091-583b-4e89-9850-cc146cef7e20","Type":"ContainerStarted","Data":"38d3f952c6da8e5abe0ff34e6f661dc813987961675c28683656ca11e0fb1956"} Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.135087 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.136844 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" event={"ID":"6b03a06a-b4d5-4443-98eb-fdea3fd8302c","Type":"ContainerStarted","Data":"e6f498ab3e3207309d95230b5295d53f036996e496406b9cb26a5ff4d9c86003"} Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.155496 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" podStartSLOduration=3.177241593 podStartE2EDuration="1m4.155474626s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.565706645 +0000 UTC m=+848.522348340" lastFinishedPulling="2025-12-02 13:19:38.543939658 +0000 UTC m=+909.500581373" observedRunningTime="2025-12-02 13:19:39.148868323 +0000 UTC m=+910.105510028" watchObservedRunningTime="2025-12-02 13:19:39.155474626 +0000 UTC m=+910.112116331" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.170574 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mg728" podStartSLOduration=7.831965889 podStartE2EDuration="17.170556308s" podCreationTimestamp="2025-12-02 13:19:22 +0000 UTC" firstStartedPulling="2025-12-02 13:19:29.040242465 +0000 UTC m=+899.996884160" lastFinishedPulling="2025-12-02 13:19:38.378832884 +0000 UTC m=+909.335474579" observedRunningTime="2025-12-02 13:19:39.163919384 +0000 UTC m=+910.120561089" watchObservedRunningTime="2025-12-02 13:19:39.170556308 +0000 UTC m=+910.127197993" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.186077 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-ctlsq" podStartSLOduration=2.14913444 podStartE2EDuration="1m3.18604597s" podCreationTimestamp="2025-12-02 13:18:36 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.576572133 +0000 UTC m=+848.533213828" lastFinishedPulling="2025-12-02 13:19:38.613483663 +0000 UTC m=+909.570125358" observedRunningTime="2025-12-02 13:19:39.182679407 +0000 UTC m=+910.139321102" watchObservedRunningTime="2025-12-02 13:19:39.18604597 +0000 UTC m=+910.142687665" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.206226 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" podStartSLOduration=3.297192604 podStartE2EDuration="1m4.206208478s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.532298191 +0000 UTC m=+848.488939886" lastFinishedPulling="2025-12-02 13:19:38.441314065 +0000 UTC m=+909.397955760" observedRunningTime="2025-12-02 13:19:39.204147017 +0000 UTC m=+910.160788712" watchObservedRunningTime="2025-12-02 13:19:39.206208478 +0000 UTC m=+910.162850173" Dec 02 13:19:39 crc kubenswrapper[4725]: I1202 13:19:39.227898 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" podStartSLOduration=3.209886489 podStartE2EDuration="1m4.227883793s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.526447396 +0000 UTC m=+848.483089091" lastFinishedPulling="2025-12-02 13:19:38.54444471 +0000 UTC m=+909.501086395" observedRunningTime="2025-12-02 13:19:39.22614224 +0000 UTC m=+910.182783935" watchObservedRunningTime="2025-12-02 13:19:39.227883793 +0000 UTC m=+910.184525478" Dec 02 13:19:41 crc kubenswrapper[4725]: I1202 13:19:41.416586 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g99db" Dec 02 13:19:42 crc kubenswrapper[4725]: I1202 13:19:42.152616 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc" Dec 02 13:19:42 crc kubenswrapper[4725]: I1202 13:19:42.620507 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:42 crc kubenswrapper[4725]: I1202 13:19:42.621530 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:42 crc kubenswrapper[4725]: I1202 13:19:42.659853 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:43 crc kubenswrapper[4725]: I1202 13:19:43.330240 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:43 crc kubenswrapper[4725]: I1202 13:19:43.394579 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mg728"] Dec 02 13:19:44 crc kubenswrapper[4725]: I1202 13:19:44.221679 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" event={"ID":"511bab3c-3885-4c5a-8c91-bd231b1c2a03","Type":"ContainerStarted","Data":"199a2966b4e15f184b2822c253bea1b0b7d75cbd8449c981c037a9bc1c840e7d"} Dec 02 13:19:44 crc kubenswrapper[4725]: I1202 13:19:44.223491 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:19:44 crc kubenswrapper[4725]: I1202 13:19:44.249522 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" podStartSLOduration=3.005468995 podStartE2EDuration="1m9.24949771s" podCreationTimestamp="2025-12-02 13:18:35 +0000 UTC" firstStartedPulling="2025-12-02 13:18:37.702939351 +0000 UTC m=+848.659581046" lastFinishedPulling="2025-12-02 13:19:43.946968066 +0000 UTC m=+914.903609761" observedRunningTime="2025-12-02 13:19:44.246086946 +0000 UTC m=+915.202728651" watchObservedRunningTime="2025-12-02 13:19:44.24949771 +0000 UTC m=+915.206139425" Dec 02 13:19:45 crc kubenswrapper[4725]: I1202 13:19:45.227326 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mg728" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="registry-server" containerID="cri-o://f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af" gracePeriod=2 Dec 02 13:19:45 crc kubenswrapper[4725]: I1202 13:19:45.890404 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-cllpp" Dec 02 13:19:45 crc kubenswrapper[4725]: I1202 13:19:45.919816 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.096038 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-catalog-content\") pod \"b7b2fb09-935a-4165-ac85-82806d1a9f10\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.096185 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-utilities\") pod \"b7b2fb09-935a-4165-ac85-82806d1a9f10\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.096245 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqhxx\" (UniqueName: \"kubernetes.io/projected/b7b2fb09-935a-4165-ac85-82806d1a9f10-kube-api-access-hqhxx\") pod \"b7b2fb09-935a-4165-ac85-82806d1a9f10\" (UID: \"b7b2fb09-935a-4165-ac85-82806d1a9f10\") " Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.096869 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-utilities" (OuterVolumeSpecName: "utilities") pod "b7b2fb09-935a-4165-ac85-82806d1a9f10" (UID: "b7b2fb09-935a-4165-ac85-82806d1a9f10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.102890 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b2fb09-935a-4165-ac85-82806d1a9f10-kube-api-access-hqhxx" (OuterVolumeSpecName: "kube-api-access-hqhxx") pod "b7b2fb09-935a-4165-ac85-82806d1a9f10" (UID: "b7b2fb09-935a-4165-ac85-82806d1a9f10"). InnerVolumeSpecName "kube-api-access-hqhxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.142973 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7b2fb09-935a-4165-ac85-82806d1a9f10" (UID: "b7b2fb09-935a-4165-ac85-82806d1a9f10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.197782 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqhxx\" (UniqueName: \"kubernetes.io/projected/b7b2fb09-935a-4165-ac85-82806d1a9f10-kube-api-access-hqhxx\") on node \"crc\" DevicePath \"\"" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.197820 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.197831 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7b2fb09-935a-4165-ac85-82806d1a9f10-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.234654 4725 generic.go:334] "Generic (PLEG): container finished" podID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerID="f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af" exitCode=0 Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.234699 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mg728" event={"ID":"b7b2fb09-935a-4165-ac85-82806d1a9f10","Type":"ContainerDied","Data":"f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af"} Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.234741 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mg728" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.234759 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mg728" event={"ID":"b7b2fb09-935a-4165-ac85-82806d1a9f10","Type":"ContainerDied","Data":"2e36ce2b0702996027ca69d66b2cef02d35b98fbefeba6d65a432220f7a59509"} Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.234779 4725 scope.go:117] "RemoveContainer" containerID="f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.251659 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.253498 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-cptsj" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.261707 4725 scope.go:117] "RemoveContainer" containerID="f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.284024 4725 scope.go:117] "RemoveContainer" containerID="6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.308593 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mg728"] Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.315601 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mg728"] Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.316489 4725 scope.go:117] "RemoveContainer" containerID="f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af" Dec 02 13:19:46 crc kubenswrapper[4725]: E1202 13:19:46.316849 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af\": container with ID starting with f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af not found: ID does not exist" containerID="f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.316881 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af"} err="failed to get container status \"f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af\": rpc error: code = NotFound desc = could not find container \"f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af\": container with ID starting with f20b330606a090c6edc20c8ca8bf63bc0a686983d17217c9b9ef64ac18d215af not found: ID does not exist" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.316901 4725 scope.go:117] "RemoveContainer" containerID="f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a" Dec 02 13:19:46 crc kubenswrapper[4725]: E1202 13:19:46.317581 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a\": container with ID starting with f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a not found: ID does not exist" containerID="f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.317602 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a"} err="failed to get container status \"f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a\": rpc error: code = NotFound desc = could not find container \"f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a\": container with ID starting with f23faa4920ab5dffe95a68097ac39b5ac9f8339cbe61fc02f0986c3cc58b385a not found: ID does not exist" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.317616 4725 scope.go:117] "RemoveContainer" containerID="6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7" Dec 02 13:19:46 crc kubenswrapper[4725]: E1202 13:19:46.318425 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7\": container with ID starting with 6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7 not found: ID does not exist" containerID="6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.318474 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7"} err="failed to get container status \"6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7\": rpc error: code = NotFound desc = could not find container \"6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7\": container with ID starting with 6c8721854053fddc43a53128a62a4215e5109e82b6563377069b51b0f6c80db7 not found: ID does not exist" Dec 02 13:19:46 crc kubenswrapper[4725]: I1202 13:19:46.461020 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-rrnmn" Dec 02 13:19:47 crc kubenswrapper[4725]: I1202 13:19:47.277253 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" path="/var/lib/kubelet/pods/b7b2fb09-935a-4165-ac85-82806d1a9f10/volumes" Dec 02 13:19:54 crc kubenswrapper[4725]: I1202 13:19:54.284453 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:19:54 crc kubenswrapper[4725]: I1202 13:19:54.284922 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:19:57 crc kubenswrapper[4725]: I1202 13:19:57.078333 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mxpqd" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.925382 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2hfzm"] Dec 02 13:20:13 crc kubenswrapper[4725]: E1202 13:20:13.926255 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="registry-server" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926268 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="registry-server" Dec 02 13:20:13 crc kubenswrapper[4725]: E1202 13:20:13.926304 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="extract-content" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926310 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="extract-content" Dec 02 13:20:13 crc kubenswrapper[4725]: E1202 13:20:13.926327 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="extract-utilities" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926334 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="extract-utilities" Dec 02 13:20:13 crc kubenswrapper[4725]: E1202 13:20:13.926342 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="extract-content" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926348 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="extract-content" Dec 02 13:20:13 crc kubenswrapper[4725]: E1202 13:20:13.926357 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="extract-utilities" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926363 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="extract-utilities" Dec 02 13:20:13 crc kubenswrapper[4725]: E1202 13:20:13.926377 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="registry-server" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926385 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="registry-server" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926544 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b2fb09-935a-4165-ac85-82806d1a9f10" containerName="registry-server" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.926562 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d64eb153-6898-4b22-b86b-df77e2c71044" containerName="registry-server" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.927273 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.931524 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wvrbf" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.934099 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.934412 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.934591 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 02 13:20:13 crc kubenswrapper[4725]: I1202 13:20:13.935756 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2hfzm"] Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.035556 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m4vnj"] Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.037020 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.047710 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.057490 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m4vnj"] Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.105922 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bb57d0-7432-4076-b850-7bbe0dca847f-config\") pod \"dnsmasq-dns-675f4bcbfc-2hfzm\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.106041 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hxm4\" (UniqueName: \"kubernetes.io/projected/f5bb57d0-7432-4076-b850-7bbe0dca847f-kube-api-access-7hxm4\") pod \"dnsmasq-dns-675f4bcbfc-2hfzm\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.207868 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hxm4\" (UniqueName: \"kubernetes.io/projected/f5bb57d0-7432-4076-b850-7bbe0dca847f-kube-api-access-7hxm4\") pod \"dnsmasq-dns-675f4bcbfc-2hfzm\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.207916 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bb57d0-7432-4076-b850-7bbe0dca847f-config\") pod \"dnsmasq-dns-675f4bcbfc-2hfzm\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.207972 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbv5b\" (UniqueName: \"kubernetes.io/projected/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-kube-api-access-cbv5b\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.208007 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.208023 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-config\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.209147 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bb57d0-7432-4076-b850-7bbe0dca847f-config\") pod \"dnsmasq-dns-675f4bcbfc-2hfzm\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.232105 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hxm4\" (UniqueName: \"kubernetes.io/projected/f5bb57d0-7432-4076-b850-7bbe0dca847f-kube-api-access-7hxm4\") pod \"dnsmasq-dns-675f4bcbfc-2hfzm\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.255843 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.310169 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbv5b\" (UniqueName: \"kubernetes.io/projected/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-kube-api-access-cbv5b\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.310230 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.310255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-config\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.311223 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-config\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.311244 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.329892 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbv5b\" (UniqueName: \"kubernetes.io/projected/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-kube-api-access-cbv5b\") pod \"dnsmasq-dns-78dd6ddcc-m4vnj\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.365750 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.837630 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2hfzm"] Dec 02 13:20:14 crc kubenswrapper[4725]: I1202 13:20:14.929071 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m4vnj"] Dec 02 13:20:14 crc kubenswrapper[4725]: W1202 13:20:14.933844 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d3bc2db_b0ba_49b9_a3a3_a0e7d93f9501.slice/crio-d6a7178cf34222285d70a43d60007160eb9c23e099a31f9b438ab10859993071 WatchSource:0}: Error finding container d6a7178cf34222285d70a43d60007160eb9c23e099a31f9b438ab10859993071: Status 404 returned error can't find the container with id d6a7178cf34222285d70a43d60007160eb9c23e099a31f9b438ab10859993071 Dec 02 13:20:15 crc kubenswrapper[4725]: I1202 13:20:15.430361 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" event={"ID":"f5bb57d0-7432-4076-b850-7bbe0dca847f","Type":"ContainerStarted","Data":"2eb8fbe2ae670cf6717e63d2cdd5220a00c2fe627407b236821405e4fc44ed5a"} Dec 02 13:20:15 crc kubenswrapper[4725]: I1202 13:20:15.431555 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" event={"ID":"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501","Type":"ContainerStarted","Data":"d6a7178cf34222285d70a43d60007160eb9c23e099a31f9b438ab10859993071"} Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.071118 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2hfzm"] Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.096799 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhxnm"] Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.098396 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.121542 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhxnm"] Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.262375 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlmcm\" (UniqueName: \"kubernetes.io/projected/0536fef7-fbd3-4633-a373-3c6e447d8a19-kube-api-access-zlmcm\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.262431 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-config\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.262517 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.364750 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlmcm\" (UniqueName: \"kubernetes.io/projected/0536fef7-fbd3-4633-a373-3c6e447d8a19-kube-api-access-zlmcm\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.365670 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-config\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.365720 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.366542 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-config\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.366698 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.393995 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m4vnj"] Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.403269 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlmcm\" (UniqueName: \"kubernetes.io/projected/0536fef7-fbd3-4633-a373-3c6e447d8a19-kube-api-access-zlmcm\") pod \"dnsmasq-dns-666b6646f7-qhxnm\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.414140 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-s7cw4"] Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.427210 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.439210 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.469211 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.469741 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rlj4\" (UniqueName: \"kubernetes.io/projected/08951dcf-7810-40b8-be36-9003fe20022b-kube-api-access-7rlj4\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.469787 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-config\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.483429 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-s7cw4"] Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.571844 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rlj4\" (UniqueName: \"kubernetes.io/projected/08951dcf-7810-40b8-be36-9003fe20022b-kube-api-access-7rlj4\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.571905 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-config\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.571988 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.573082 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.574405 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-config\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.595602 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rlj4\" (UniqueName: \"kubernetes.io/projected/08951dcf-7810-40b8-be36-9003fe20022b-kube-api-access-7rlj4\") pod \"dnsmasq-dns-57d769cc4f-s7cw4\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:17 crc kubenswrapper[4725]: I1202 13:20:17.840920 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.140105 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhxnm"] Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.269427 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.274073 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.278578 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hh5j8" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.278825 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.279052 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.279227 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.279368 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.279546 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.279623 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.296332 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415290 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415600 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415656 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-config-data\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415673 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cln48\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-kube-api-access-cln48\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415692 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415722 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415763 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d630b03-f4dd-48ac-a72b-b457e3651c76-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415779 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415811 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415827 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d630b03-f4dd-48ac-a72b-b457e3651c76-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.415888 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.483216 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" event={"ID":"0536fef7-fbd3-4633-a373-3c6e447d8a19","Type":"ContainerStarted","Data":"3e843a4ab2f8dd720336d5d8934cfaaf2f3f22aba440a0fc486199ccdc0baf2e"} Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.496630 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-s7cw4"] Dec 02 13:20:18 crc kubenswrapper[4725]: W1202 13:20:18.504436 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08951dcf_7810_40b8_be36_9003fe20022b.slice/crio-811189f07385b84b3ebff449dd93e8dbf38172cf04d1936489802c288c4adfb7 WatchSource:0}: Error finding container 811189f07385b84b3ebff449dd93e8dbf38172cf04d1936489802c288c4adfb7: Status 404 returned error can't find the container with id 811189f07385b84b3ebff449dd93e8dbf38172cf04d1936489802c288c4adfb7 Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517451 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-config-data\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517515 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cln48\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-kube-api-access-cln48\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517552 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517575 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517622 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517644 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d630b03-f4dd-48ac-a72b-b457e3651c76-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517667 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517684 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d630b03-f4dd-48ac-a72b-b457e3651c76-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517750 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517776 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.517806 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.518203 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.519096 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-config-data\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.520122 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.520254 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.521378 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.523990 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.526914 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d630b03-f4dd-48ac-a72b-b457e3651c76-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.526953 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d630b03-f4dd-48ac-a72b-b457e3651c76-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.529103 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.530752 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.545324 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.552276 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.553131 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cln48\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-kube-api-access-cln48\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.559921 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.561443 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.561575 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.561743 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-n5vfx" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.561791 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.561945 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.562108 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.562292 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.572314 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.609864 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618764 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618817 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ede4bdcc-2588-4e91-80f7-1bcf0467e861-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618884 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618912 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4c2m\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-kube-api-access-c4c2m\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618927 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ede4bdcc-2588-4e91-80f7-1bcf0467e861-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618941 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618967 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.618990 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.619013 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.619043 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.619062 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720447 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720559 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4c2m\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-kube-api-access-c4c2m\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720611 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ede4bdcc-2588-4e91-80f7-1bcf0467e861-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720629 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720658 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720681 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720700 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720725 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720744 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720782 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.720804 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ede4bdcc-2588-4e91-80f7-1bcf0467e861-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.722217 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.722312 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.722426 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.722740 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.722889 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.722958 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.730911 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.731665 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ede4bdcc-2588-4e91-80f7-1bcf0467e861-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.731848 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.739908 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ede4bdcc-2588-4e91-80f7-1bcf0467e861-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.747818 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4c2m\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-kube-api-access-c4c2m\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.762439 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:18 crc kubenswrapper[4725]: I1202 13:20:18.927719 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.031560 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.544735 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d630b03-f4dd-48ac-a72b-b457e3651c76","Type":"ContainerStarted","Data":"f0fffedbd155c725d59b39b63b9c13da0a5b9651c2b1e461e1d4a301a2b91d94"} Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.561523 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" event={"ID":"08951dcf-7810-40b8-be36-9003fe20022b","Type":"ContainerStarted","Data":"811189f07385b84b3ebff449dd93e8dbf38172cf04d1936489802c288c4adfb7"} Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.731447 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:20:19 crc kubenswrapper[4725]: W1202 13:20:19.767542 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podede4bdcc_2588_4e91_80f7_1bcf0467e861.slice/crio-50642c4080773e0a09deda9800dcf1fb03d37093730302089399718e960cd355 WatchSource:0}: Error finding container 50642c4080773e0a09deda9800dcf1fb03d37093730302089399718e960cd355: Status 404 returned error can't find the container with id 50642c4080773e0a09deda9800dcf1fb03d37093730302089399718e960cd355 Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.803675 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.805402 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.811647 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.811669 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.811648 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.811984 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mvxgq" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.818853 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.822621 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.965797 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4wg4\" (UniqueName: \"kubernetes.io/projected/ef9cb921-2581-495e-be27-cae739b9a971-kube-api-access-x4wg4\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.965903 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef9cb921-2581-495e-be27-cae739b9a971-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.965940 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9cb921-2581-495e-be27-cae739b9a971-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.966029 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.966063 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.966100 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ef9cb921-2581-495e-be27-cae739b9a971-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.966140 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-config-data-default\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:19 crc kubenswrapper[4725]: I1202 13:20:19.966183 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-kolla-config\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.067665 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ef9cb921-2581-495e-be27-cae739b9a971-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.067741 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-config-data-default\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.067780 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-kolla-config\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.067816 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4wg4\" (UniqueName: \"kubernetes.io/projected/ef9cb921-2581-495e-be27-cae739b9a971-kube-api-access-x4wg4\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.067864 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef9cb921-2581-495e-be27-cae739b9a971-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.067891 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9cb921-2581-495e-be27-cae739b9a971-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.068078 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.068235 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ef9cb921-2581-495e-be27-cae739b9a971-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.068147 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.069085 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.069426 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-config-data-default\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.070230 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-kolla-config\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.083957 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef9cb921-2581-495e-be27-cae739b9a971-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.090742 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9cb921-2581-495e-be27-cae739b9a971-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.091241 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef9cb921-2581-495e-be27-cae739b9a971-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.091977 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4wg4\" (UniqueName: \"kubernetes.io/projected/ef9cb921-2581-495e-be27-cae739b9a971-kube-api-access-x4wg4\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.092674 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"ef9cb921-2581-495e-be27-cae739b9a971\") " pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.132139 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.572871 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ede4bdcc-2588-4e91-80f7-1bcf0467e861","Type":"ContainerStarted","Data":"50642c4080773e0a09deda9800dcf1fb03d37093730302089399718e960cd355"} Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.996621 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 13:20:20 crc kubenswrapper[4725]: I1202 13:20:20.998319 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.000268 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.008518 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hb42k" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.008750 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.008882 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.025596 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097337 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc289673-6c5c-45f0-a9ff-df10cddd635c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097418 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cc289673-6c5c-45f0-a9ff-df10cddd635c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097443 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097521 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097668 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097774 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097901 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gq2t\" (UniqueName: \"kubernetes.io/projected/cc289673-6c5c-45f0-a9ff-df10cddd635c-kube-api-access-7gq2t\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.097927 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc289673-6c5c-45f0-a9ff-df10cddd635c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.156766 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.200814 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc289673-6c5c-45f0-a9ff-df10cddd635c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201086 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cc289673-6c5c-45f0-a9ff-df10cddd635c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201108 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201138 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201162 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201202 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gq2t\" (UniqueName: \"kubernetes.io/projected/cc289673-6c5c-45f0-a9ff-df10cddd635c-kube-api-access-7gq2t\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.201272 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc289673-6c5c-45f0-a9ff-df10cddd635c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.202320 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.203312 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.205675 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.205996 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cc289673-6c5c-45f0-a9ff-df10cddd635c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.207342 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc289673-6c5c-45f0-a9ff-df10cddd635c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.210051 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc289673-6c5c-45f0-a9ff-df10cddd635c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.210417 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cc289673-6c5c-45f0-a9ff-df10cddd635c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.227041 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gq2t\" (UniqueName: \"kubernetes.io/projected/cc289673-6c5c-45f0-a9ff-df10cddd635c-kube-api-access-7gq2t\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.249637 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"cc289673-6c5c-45f0-a9ff-df10cddd635c\") " pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.346011 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.362342 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.363305 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.373185 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-rhtlz" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.373310 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.373394 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.384889 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.445901 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbf393-00ed-46c6-a61f-8a458212e8e7-config-data\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.445968 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/21bbf393-00ed-46c6-a61f-8a458212e8e7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.446017 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/21bbf393-00ed-46c6-a61f-8a458212e8e7-kolla-config\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.446062 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5qbl\" (UniqueName: \"kubernetes.io/projected/21bbf393-00ed-46c6-a61f-8a458212e8e7-kube-api-access-b5qbl\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.446166 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21bbf393-00ed-46c6-a61f-8a458212e8e7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.547062 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbf393-00ed-46c6-a61f-8a458212e8e7-config-data\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.547362 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/21bbf393-00ed-46c6-a61f-8a458212e8e7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.547390 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/21bbf393-00ed-46c6-a61f-8a458212e8e7-kolla-config\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.547415 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5qbl\" (UniqueName: \"kubernetes.io/projected/21bbf393-00ed-46c6-a61f-8a458212e8e7-kube-api-access-b5qbl\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.547453 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21bbf393-00ed-46c6-a61f-8a458212e8e7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.548100 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/21bbf393-00ed-46c6-a61f-8a458212e8e7-config-data\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.548667 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/21bbf393-00ed-46c6-a61f-8a458212e8e7-kolla-config\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.557756 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21bbf393-00ed-46c6-a61f-8a458212e8e7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.559353 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/21bbf393-00ed-46c6-a61f-8a458212e8e7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.569454 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5qbl\" (UniqueName: \"kubernetes.io/projected/21bbf393-00ed-46c6-a61f-8a458212e8e7-kube-api-access-b5qbl\") pod \"memcached-0\" (UID: \"21bbf393-00ed-46c6-a61f-8a458212e8e7\") " pod="openstack/memcached-0" Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.591175 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ef9cb921-2581-495e-be27-cae739b9a971","Type":"ContainerStarted","Data":"86420bfbbdd1077c2a149001697909819e212ef9f37c7b30b5892ca488dac5be"} Dec 02 13:20:21 crc kubenswrapper[4725]: I1202 13:20:21.710594 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 02 13:20:22 crc kubenswrapper[4725]: I1202 13:20:22.227288 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.120226 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.121715 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.124628 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7qxzw" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.132721 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.234265 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfszl\" (UniqueName: \"kubernetes.io/projected/1bbe1936-05fb-4f72-affd-7573be426c6e-kube-api-access-gfszl\") pod \"kube-state-metrics-0\" (UID: \"1bbe1936-05fb-4f72-affd-7573be426c6e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.285056 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.285412 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.285494 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.286362 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"203c78b7d89a690dfa6df4df74f2999e81c6261249dd7853e1b1c46205cb6197"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.286425 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://203c78b7d89a690dfa6df4df74f2999e81c6261249dd7853e1b1c46205cb6197" gracePeriod=600 Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.335332 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfszl\" (UniqueName: \"kubernetes.io/projected/1bbe1936-05fb-4f72-affd-7573be426c6e-kube-api-access-gfszl\") pod \"kube-state-metrics-0\" (UID: \"1bbe1936-05fb-4f72-affd-7573be426c6e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.353221 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfszl\" (UniqueName: \"kubernetes.io/projected/1bbe1936-05fb-4f72-affd-7573be426c6e-kube-api-access-gfszl\") pod \"kube-state-metrics-0\" (UID: \"1bbe1936-05fb-4f72-affd-7573be426c6e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.458938 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.626043 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="203c78b7d89a690dfa6df4df74f2999e81c6261249dd7853e1b1c46205cb6197" exitCode=0 Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.626115 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"203c78b7d89a690dfa6df4df74f2999e81c6261249dd7853e1b1c46205cb6197"} Dec 02 13:20:24 crc kubenswrapper[4725]: I1202 13:20:24.626190 4725 scope.go:117] "RemoveContainer" containerID="01e784ee70283b45e319c7d0015626dcb847bd5037e05305c113f2ede968d762" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.713644 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ngrg"] Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.715697 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.718232 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.718304 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.725677 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-r8f7d" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.739665 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ngrg"] Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.778065 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-bvlkl"] Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.780182 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.794214 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bvlkl"] Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.817492 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-log-ovn\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.817777 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-scripts\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.817906 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-log\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818109 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pmd6\" (UniqueName: \"kubernetes.io/projected/c552e1c0-2902-4b74-800f-5bccc6ba4427-kube-api-access-6pmd6\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818172 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8kq\" (UniqueName: \"kubernetes.io/projected/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-kube-api-access-tc8kq\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818218 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-ovn-controller-tls-certs\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818322 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-lib\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818350 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-run\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818364 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c552e1c0-2902-4b74-800f-5bccc6ba4427-scripts\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818444 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-combined-ca-bundle\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818495 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-run-ovn\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818522 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-run\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.818551 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-etc-ovs\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.920338 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pmd6\" (UniqueName: \"kubernetes.io/projected/c552e1c0-2902-4b74-800f-5bccc6ba4427-kube-api-access-6pmd6\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.920711 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8kq\" (UniqueName: \"kubernetes.io/projected/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-kube-api-access-tc8kq\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.920891 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-ovn-controller-tls-certs\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.920955 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-lib\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.920995 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-run\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921012 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c552e1c0-2902-4b74-800f-5bccc6ba4427-scripts\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921400 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-run\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921477 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-combined-ca-bundle\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921740 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-run-ovn\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921506 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-run-ovn\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921878 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-run\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921906 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-etc-ovs\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921932 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-log-ovn\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921953 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-scripts\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:27 crc kubenswrapper[4725]: I1202 13:20:27.921998 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-log\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.921995 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-run\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.922187 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-var-log-ovn\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.922197 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-etc-ovs\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.922337 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-log\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.923651 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c552e1c0-2902-4b74-800f-5bccc6ba4427-scripts\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.923786 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c552e1c0-2902-4b74-800f-5bccc6ba4427-var-lib\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.924232 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-scripts\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.925909 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-combined-ca-bundle\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.933032 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-ovn-controller-tls-certs\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.957003 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8kq\" (UniqueName: \"kubernetes.io/projected/79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7-kube-api-access-tc8kq\") pod \"ovn-controller-4ngrg\" (UID: \"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7\") " pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:27.964973 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pmd6\" (UniqueName: \"kubernetes.io/projected/c552e1c0-2902-4b74-800f-5bccc6ba4427-kube-api-access-6pmd6\") pod \"ovn-controller-ovs-bvlkl\" (UID: \"c552e1c0-2902-4b74-800f-5bccc6ba4427\") " pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.038867 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.100062 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.338786 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.349130 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.352798 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.353172 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.355662 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.355758 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.368669 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mcvsg" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.371732 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439624 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bad22ab8-0033-4a28-b7aa-65b1c285a989-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439675 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bad22ab8-0033-4a28-b7aa-65b1c285a989-config\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439704 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439846 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439893 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bad22ab8-0033-4a28-b7aa-65b1c285a989-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439923 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gpss\" (UniqueName: \"kubernetes.io/projected/bad22ab8-0033-4a28-b7aa-65b1c285a989-kube-api-access-9gpss\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439960 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.439994 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541542 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bad22ab8-0033-4a28-b7aa-65b1c285a989-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541582 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bad22ab8-0033-4a28-b7aa-65b1c285a989-config\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541612 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541714 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541746 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bad22ab8-0033-4a28-b7aa-65b1c285a989-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541774 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gpss\" (UniqueName: \"kubernetes.io/projected/bad22ab8-0033-4a28-b7aa-65b1c285a989-kube-api-access-9gpss\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541809 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.541845 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.542395 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.544135 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bad22ab8-0033-4a28-b7aa-65b1c285a989-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.545895 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bad22ab8-0033-4a28-b7aa-65b1c285a989-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.546591 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bad22ab8-0033-4a28-b7aa-65b1c285a989-config\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.551827 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.551937 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.552968 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bad22ab8-0033-4a28-b7aa-65b1c285a989-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.561402 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gpss\" (UniqueName: \"kubernetes.io/projected/bad22ab8-0033-4a28-b7aa-65b1c285a989-kube-api-access-9gpss\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.570945 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bad22ab8-0033-4a28-b7aa-65b1c285a989\") " pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.681481 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 02 13:20:28 crc kubenswrapper[4725]: I1202 13:20:28.715534 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc289673-6c5c-45f0-a9ff-df10cddd635c","Type":"ContainerStarted","Data":"9e4e0e01eec0e8924725834e4930ccdceff00ba5409fc1cd293f7713e362f096"} Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.481772 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.485602 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.488643 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-zfjx2" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.489052 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.489225 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.490673 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.503625 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505160 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-745xk\" (UniqueName: \"kubernetes.io/projected/51144a39-c6dd-4a5f-adfb-6d62cb459485-kube-api-access-745xk\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505196 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51144a39-c6dd-4a5f-adfb-6d62cb459485-config\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505241 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51144a39-c6dd-4a5f-adfb-6d62cb459485-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505262 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505287 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505320 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505352 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.505376 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51144a39-c6dd-4a5f-adfb-6d62cb459485-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.607617 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-745xk\" (UniqueName: \"kubernetes.io/projected/51144a39-c6dd-4a5f-adfb-6d62cb459485-kube-api-access-745xk\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.609982 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51144a39-c6dd-4a5f-adfb-6d62cb459485-config\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.610050 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51144a39-c6dd-4a5f-adfb-6d62cb459485-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.610083 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.610114 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.610161 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.610202 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.610237 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51144a39-c6dd-4a5f-adfb-6d62cb459485-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.611736 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.612350 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51144a39-c6dd-4a5f-adfb-6d62cb459485-config\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.615328 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51144a39-c6dd-4a5f-adfb-6d62cb459485-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.618700 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.625551 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-745xk\" (UniqueName: \"kubernetes.io/projected/51144a39-c6dd-4a5f-adfb-6d62cb459485-kube-api-access-745xk\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.626892 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51144a39-c6dd-4a5f-adfb-6d62cb459485-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.631445 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.631529 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51144a39-c6dd-4a5f-adfb-6d62cb459485-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.647851 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-sb-0\" (UID: \"51144a39-c6dd-4a5f-adfb-6d62cb459485\") " pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:31 crc kubenswrapper[4725]: I1202 13:20:31.814840 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.173445 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.174176 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cln48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(0d630b03-f4dd-48ac-a72b-b457e3651c76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.175388 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.844040 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.926812 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.927641 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cbv5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-m4vnj_openstack(8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:20:40 crc kubenswrapper[4725]: E1202 13:20:40.928986 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" podUID="8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501" Dec 02 13:20:42 crc kubenswrapper[4725]: E1202 13:20:42.759307 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 13:20:42 crc kubenswrapper[4725]: E1202 13:20:42.759692 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7hxm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-2hfzm_openstack(f5bb57d0-7432-4076-b850-7bbe0dca847f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:20:42 crc kubenswrapper[4725]: E1202 13:20:42.760856 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" podUID="f5bb57d0-7432-4076-b850-7bbe0dca847f" Dec 02 13:20:42 crc kubenswrapper[4725]: E1202 13:20:42.822442 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 13:20:42 crc kubenswrapper[4725]: E1202 13:20:42.823348 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zlmcm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-qhxnm_openstack(0536fef7-fbd3-4633-a373-3c6e447d8a19): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:20:42 crc kubenswrapper[4725]: E1202 13:20:42.824580 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" Dec 02 13:20:42 crc kubenswrapper[4725]: I1202 13:20:42.887320 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" event={"ID":"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501","Type":"ContainerDied","Data":"d6a7178cf34222285d70a43d60007160eb9c23e099a31f9b438ab10859993071"} Dec 02 13:20:42 crc kubenswrapper[4725]: I1202 13:20:42.887360 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6a7178cf34222285d70a43d60007160eb9c23e099a31f9b438ab10859993071" Dec 02 13:20:42 crc kubenswrapper[4725]: I1202 13:20:42.890015 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.015520 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbv5b\" (UniqueName: \"kubernetes.io/projected/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-kube-api-access-cbv5b\") pod \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.015921 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-dns-svc\") pod \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.016019 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-config\") pod \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\" (UID: \"8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501\") " Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.016780 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501" (UID: "8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.017146 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-config" (OuterVolumeSpecName: "config") pod "8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501" (UID: "8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.022324 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-kube-api-access-cbv5b" (OuterVolumeSpecName: "kube-api-access-cbv5b") pod "8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501" (UID: "8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501"). InnerVolumeSpecName "kube-api-access-cbv5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.117738 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbv5b\" (UniqueName: \"kubernetes.io/projected/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-kube-api-access-cbv5b\") on node \"crc\" DevicePath \"\"" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.117981 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.117994 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.428986 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.454239 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.455902 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.525387 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hxm4\" (UniqueName: \"kubernetes.io/projected/f5bb57d0-7432-4076-b850-7bbe0dca847f-kube-api-access-7hxm4\") pod \"f5bb57d0-7432-4076-b850-7bbe0dca847f\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.525525 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bb57d0-7432-4076-b850-7bbe0dca847f-config\") pod \"f5bb57d0-7432-4076-b850-7bbe0dca847f\" (UID: \"f5bb57d0-7432-4076-b850-7bbe0dca847f\") " Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.526846 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5bb57d0-7432-4076-b850-7bbe0dca847f-config" (OuterVolumeSpecName: "config") pod "f5bb57d0-7432-4076-b850-7bbe0dca847f" (UID: "f5bb57d0-7432-4076-b850-7bbe0dca847f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.535621 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5bb57d0-7432-4076-b850-7bbe0dca847f-kube-api-access-7hxm4" (OuterVolumeSpecName: "kube-api-access-7hxm4") pod "f5bb57d0-7432-4076-b850-7bbe0dca847f" (UID: "f5bb57d0-7432-4076-b850-7bbe0dca847f"). InnerVolumeSpecName "kube-api-access-7hxm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.627474 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bb57d0-7432-4076-b850-7bbe0dca847f-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.627911 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hxm4\" (UniqueName: \"kubernetes.io/projected/f5bb57d0-7432-4076-b850-7bbe0dca847f-kube-api-access-7hxm4\") on node \"crc\" DevicePath \"\"" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.829255 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bvlkl"] Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.837082 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ngrg"] Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.901822 4725 generic.go:334] "Generic (PLEG): container finished" podID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerID="936a3e684329aa1d63ab9776d68a76e8deaba48749057b4e2377a8bc56c0a6bc" exitCode=0 Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.901910 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" event={"ID":"0536fef7-fbd3-4633-a373-3c6e447d8a19","Type":"ContainerDied","Data":"936a3e684329aa1d63ab9776d68a76e8deaba48749057b4e2377a8bc56c0a6bc"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.903759 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ngrg" event={"ID":"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7","Type":"ContainerStarted","Data":"5abfb40d23576746d4775db367e8c6d9325a747899c8e4d3b4481324542ecb15"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.905310 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"21bbf393-00ed-46c6-a61f-8a458212e8e7","Type":"ContainerStarted","Data":"362a7c99bf964d880291a915f70e6d3a2d60263fa6345d135fd112cf2d5d77b7"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.906722 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bbe1936-05fb-4f72-affd-7573be426c6e","Type":"ContainerStarted","Data":"6f042378bed374a085970210d8d4b7886ba945df28b7a2defd98e13928b6e590"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.907073 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.908444 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" event={"ID":"f5bb57d0-7432-4076-b850-7bbe0dca847f","Type":"ContainerDied","Data":"2eb8fbe2ae670cf6717e63d2cdd5220a00c2fe627407b236821405e4fc44ed5a"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.908512 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2hfzm" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.910979 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc289673-6c5c-45f0-a9ff-df10cddd635c","Type":"ContainerStarted","Data":"6951fe344d501bb92edf38cef32de719f3978ddc440444e32f4029f84817d672"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.913812 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"1cf7371c50f3e4f7f952b1130b75faf18d2e6023f1c852348322f64d11754fac"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.915446 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ef9cb921-2581-495e-be27-cae739b9a971","Type":"ContainerStarted","Data":"e422d339f2cd28a78e02046fcb24608583abed1273e720204a09a1d96e906501"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.917023 4725 generic.go:334] "Generic (PLEG): container finished" podID="08951dcf-7810-40b8-be36-9003fe20022b" containerID="73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce" exitCode=0 Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.917078 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" event={"ID":"08951dcf-7810-40b8-be36-9003fe20022b","Type":"ContainerDied","Data":"73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce"} Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.919546 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-m4vnj" Dec 02 13:20:43 crc kubenswrapper[4725]: I1202 13:20:43.919769 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bvlkl" event={"ID":"c552e1c0-2902-4b74-800f-5bccc6ba4427","Type":"ContainerStarted","Data":"91b17b4882efa20a9448958c1209725852e62c7e238b133991adb4ecf4628954"} Dec 02 13:20:44 crc kubenswrapper[4725]: W1202 13:20:44.002957 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbad22ab8_0033_4a28_b7aa_65b1c285a989.slice/crio-b42787b0fb80e8989b556c306e08b688b8e31cfa003b265023abc22e98702455 WatchSource:0}: Error finding container b42787b0fb80e8989b556c306e08b688b8e31cfa003b265023abc22e98702455: Status 404 returned error can't find the container with id b42787b0fb80e8989b556c306e08b688b8e31cfa003b265023abc22e98702455 Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.018259 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.062232 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2hfzm"] Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.068401 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2hfzm"] Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.098862 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m4vnj"] Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.104328 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-m4vnj"] Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.926786 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bad22ab8-0033-4a28-b7aa-65b1c285a989","Type":"ContainerStarted","Data":"b42787b0fb80e8989b556c306e08b688b8e31cfa003b265023abc22e98702455"} Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.929169 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"51144a39-c6dd-4a5f-adfb-6d62cb459485","Type":"ContainerStarted","Data":"92bf4bd2295bcac9196318c9fa5dc716267e90e90288e2d0a73c06a1d29ead70"} Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.930937 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ede4bdcc-2588-4e91-80f7-1bcf0467e861","Type":"ContainerStarted","Data":"6e3ec9fe8664b961c59b197cc682fd6ee84f663d3385bbcb3233ec98fc495c03"} Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.933078 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" event={"ID":"0536fef7-fbd3-4633-a373-3c6e447d8a19","Type":"ContainerStarted","Data":"c4b8d491ed83ce2e962346bee5c24b02e7af38bb58e46c51daf0a8ca741b9f5e"} Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.933235 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.935707 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" event={"ID":"08951dcf-7810-40b8-be36-9003fe20022b","Type":"ContainerStarted","Data":"100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df"} Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.977059 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" podStartSLOduration=3.642412711 podStartE2EDuration="27.977040844s" podCreationTimestamp="2025-12-02 13:20:17 +0000 UTC" firstStartedPulling="2025-12-02 13:20:18.508001652 +0000 UTC m=+949.464643347" lastFinishedPulling="2025-12-02 13:20:42.842629785 +0000 UTC m=+973.799271480" observedRunningTime="2025-12-02 13:20:44.975932276 +0000 UTC m=+975.932573971" watchObservedRunningTime="2025-12-02 13:20:44.977040844 +0000 UTC m=+975.933682539" Dec 02 13:20:44 crc kubenswrapper[4725]: I1202 13:20:44.991545 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" podStartSLOduration=-9223372008.863249 podStartE2EDuration="27.991526454s" podCreationTimestamp="2025-12-02 13:20:17 +0000 UTC" firstStartedPulling="2025-12-02 13:20:18.159830272 +0000 UTC m=+949.116471967" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:20:44.988662392 +0000 UTC m=+975.945304097" watchObservedRunningTime="2025-12-02 13:20:44.991526454 +0000 UTC m=+975.948168149" Dec 02 13:20:45 crc kubenswrapper[4725]: I1202 13:20:45.282087 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501" path="/var/lib/kubelet/pods/8d3bc2db-b0ba-49b9-a3a3-a0e7d93f9501/volumes" Dec 02 13:20:45 crc kubenswrapper[4725]: I1202 13:20:45.282771 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5bb57d0-7432-4076-b850-7bbe0dca847f" path="/var/lib/kubelet/pods/f5bb57d0-7432-4076-b850-7bbe0dca847f/volumes" Dec 02 13:20:45 crc kubenswrapper[4725]: I1202 13:20:45.943062 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:47 crc kubenswrapper[4725]: I1202 13:20:47.958277 4725 generic.go:334] "Generic (PLEG): container finished" podID="cc289673-6c5c-45f0-a9ff-df10cddd635c" containerID="6951fe344d501bb92edf38cef32de719f3978ddc440444e32f4029f84817d672" exitCode=0 Dec 02 13:20:47 crc kubenswrapper[4725]: I1202 13:20:47.958642 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc289673-6c5c-45f0-a9ff-df10cddd635c","Type":"ContainerDied","Data":"6951fe344d501bb92edf38cef32de719f3978ddc440444e32f4029f84817d672"} Dec 02 13:20:47 crc kubenswrapper[4725]: I1202 13:20:47.969885 4725 generic.go:334] "Generic (PLEG): container finished" podID="ef9cb921-2581-495e-be27-cae739b9a971" containerID="e422d339f2cd28a78e02046fcb24608583abed1273e720204a09a1d96e906501" exitCode=0 Dec 02 13:20:47 crc kubenswrapper[4725]: I1202 13:20:47.969953 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ef9cb921-2581-495e-be27-cae739b9a971","Type":"ContainerDied","Data":"e422d339f2cd28a78e02046fcb24608583abed1273e720204a09a1d96e906501"} Dec 02 13:20:48 crc kubenswrapper[4725]: I1202 13:20:48.985306 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bbe1936-05fb-4f72-affd-7573be426c6e","Type":"ContainerStarted","Data":"a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1"} Dec 02 13:20:48 crc kubenswrapper[4725]: I1202 13:20:48.985912 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 02 13:20:48 crc kubenswrapper[4725]: I1202 13:20:48.990114 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"21bbf393-00ed-46c6-a61f-8a458212e8e7","Type":"ContainerStarted","Data":"79839038bbad413888cef82ee31cf0c83c532c9298724445124c82ef21d21e80"} Dec 02 13:20:48 crc kubenswrapper[4725]: I1202 13:20:48.990811 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 02 13:20:48 crc kubenswrapper[4725]: I1202 13:20:48.994762 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cc289673-6c5c-45f0-a9ff-df10cddd635c","Type":"ContainerStarted","Data":"9accaa3ec6e1eb3bdbf5f221cfad44c399ce9a2ec612d7d54f333cd242e956ae"} Dec 02 13:20:48 crc kubenswrapper[4725]: I1202 13:20:48.997200 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"51144a39-c6dd-4a5f-adfb-6d62cb459485","Type":"ContainerStarted","Data":"f9261dc2e52deeb185df67a49eb5dd491c3f252929ceeaabc8dd9a5f1a7258e5"} Dec 02 13:20:49 crc kubenswrapper[4725]: I1202 13:20:49.004006 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.832280104 podStartE2EDuration="25.003992806s" podCreationTimestamp="2025-12-02 13:20:24 +0000 UTC" firstStartedPulling="2025-12-02 13:20:43.438022805 +0000 UTC m=+974.394664500" lastFinishedPulling="2025-12-02 13:20:48.609735507 +0000 UTC m=+979.566377202" observedRunningTime="2025-12-02 13:20:48.999270149 +0000 UTC m=+979.955911874" watchObservedRunningTime="2025-12-02 13:20:49.003992806 +0000 UTC m=+979.960634501" Dec 02 13:20:49 crc kubenswrapper[4725]: I1202 13:20:49.004941 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ef9cb921-2581-495e-be27-cae739b9a971","Type":"ContainerStarted","Data":"420537540c7e2ef6636d431fae70823ec03dbad018784e5f5b495765b5383bf6"} Dec 02 13:20:49 crc kubenswrapper[4725]: I1202 13:20:49.010244 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bad22ab8-0033-4a28-b7aa-65b1c285a989","Type":"ContainerStarted","Data":"4d992ac574445c0151f6b2a2b8fd98435630f67803b06b3925794223a1e6236a"} Dec 02 13:20:49 crc kubenswrapper[4725]: I1202 13:20:49.028874 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.880376884 podStartE2EDuration="28.028858273s" podCreationTimestamp="2025-12-02 13:20:21 +0000 UTC" firstStartedPulling="2025-12-02 13:20:43.461281789 +0000 UTC m=+974.417923484" lastFinishedPulling="2025-12-02 13:20:48.609763178 +0000 UTC m=+979.566404873" observedRunningTime="2025-12-02 13:20:49.023130581 +0000 UTC m=+979.979772276" watchObservedRunningTime="2025-12-02 13:20:49.028858273 +0000 UTC m=+979.985499968" Dec 02 13:20:49 crc kubenswrapper[4725]: I1202 13:20:49.047126 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=15.675430807 podStartE2EDuration="30.047108286s" podCreationTimestamp="2025-12-02 13:20:19 +0000 UTC" firstStartedPulling="2025-12-02 13:20:28.542129522 +0000 UTC m=+959.498771217" lastFinishedPulling="2025-12-02 13:20:42.913807001 +0000 UTC m=+973.870448696" observedRunningTime="2025-12-02 13:20:49.044066281 +0000 UTC m=+980.000707976" watchObservedRunningTime="2025-12-02 13:20:49.047108286 +0000 UTC m=+980.003749981" Dec 02 13:20:49 crc kubenswrapper[4725]: I1202 13:20:49.069301 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.446898182 podStartE2EDuration="31.069261336s" podCreationTimestamp="2025-12-02 13:20:18 +0000 UTC" firstStartedPulling="2025-12-02 13:20:21.205361103 +0000 UTC m=+952.162002798" lastFinishedPulling="2025-12-02 13:20:42.827724247 +0000 UTC m=+973.784365952" observedRunningTime="2025-12-02 13:20:49.063388321 +0000 UTC m=+980.020030016" watchObservedRunningTime="2025-12-02 13:20:49.069261336 +0000 UTC m=+980.025903021" Dec 02 13:20:50 crc kubenswrapper[4725]: I1202 13:20:50.018296 4725 generic.go:334] "Generic (PLEG): container finished" podID="c552e1c0-2902-4b74-800f-5bccc6ba4427" containerID="276565115d10e22e4b0edaacbd43031b3b96faa53b75482dfa8ad9e56f3a08dc" exitCode=0 Dec 02 13:20:50 crc kubenswrapper[4725]: I1202 13:20:50.018446 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bvlkl" event={"ID":"c552e1c0-2902-4b74-800f-5bccc6ba4427","Type":"ContainerDied","Data":"276565115d10e22e4b0edaacbd43031b3b96faa53b75482dfa8ad9e56f3a08dc"} Dec 02 13:20:50 crc kubenswrapper[4725]: I1202 13:20:50.021066 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ngrg" event={"ID":"79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7","Type":"ContainerStarted","Data":"882f3405ad4222125126f3d986832e91d00720a15c901b31000ce7722fcffcde"} Dec 02 13:20:50 crc kubenswrapper[4725]: I1202 13:20:50.059715 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4ngrg" podStartSLOduration=18.169744591 podStartE2EDuration="23.059695396s" podCreationTimestamp="2025-12-02 13:20:27 +0000 UTC" firstStartedPulling="2025-12-02 13:20:43.808384782 +0000 UTC m=+974.765026477" lastFinishedPulling="2025-12-02 13:20:48.698335597 +0000 UTC m=+979.654977282" observedRunningTime="2025-12-02 13:20:50.054493577 +0000 UTC m=+981.011135262" watchObservedRunningTime="2025-12-02 13:20:50.059695396 +0000 UTC m=+981.016337091" Dec 02 13:20:50 crc kubenswrapper[4725]: I1202 13:20:50.139493 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 02 13:20:50 crc kubenswrapper[4725]: I1202 13:20:50.139796 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 02 13:20:51 crc kubenswrapper[4725]: I1202 13:20:51.029760 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bvlkl" event={"ID":"c552e1c0-2902-4b74-800f-5bccc6ba4427","Type":"ContainerStarted","Data":"4c494c20a120e56463fcccd1b48ccedd878241a524bcc263e710647b58fd0232"} Dec 02 13:20:51 crc kubenswrapper[4725]: I1202 13:20:51.030079 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4ngrg" Dec 02 13:20:51 crc kubenswrapper[4725]: I1202 13:20:51.350509 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:51 crc kubenswrapper[4725]: I1202 13:20:51.351306 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 02 13:20:52 crc kubenswrapper[4725]: I1202 13:20:52.429692 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:20:52 crc kubenswrapper[4725]: I1202 13:20:52.844652 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:20:52 crc kubenswrapper[4725]: I1202 13:20:52.901814 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhxnm"] Dec 02 13:20:53 crc kubenswrapper[4725]: I1202 13:20:53.677553 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="dnsmasq-dns" containerID="cri-o://c4b8d491ed83ce2e962346bee5c24b02e7af38bb58e46c51daf0a8ca741b9f5e" gracePeriod=10 Dec 02 13:20:54 crc kubenswrapper[4725]: I1202 13:20:54.464745 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.716933 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.823270 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d630b03-f4dd-48ac-a72b-b457e3651c76","Type":"ContainerStarted","Data":"988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113"} Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.850422 4725 generic.go:334] "Generic (PLEG): container finished" podID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerID="c4b8d491ed83ce2e962346bee5c24b02e7af38bb58e46c51daf0a8ca741b9f5e" exitCode=0 Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.850865 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" event={"ID":"0536fef7-fbd3-4633-a373-3c6e447d8a19","Type":"ContainerDied","Data":"c4b8d491ed83ce2e962346bee5c24b02e7af38bb58e46c51daf0a8ca741b9f5e"} Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.854275 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bvlkl" event={"ID":"c552e1c0-2902-4b74-800f-5bccc6ba4427","Type":"ContainerStarted","Data":"97de859e305a72b8e70c452a047087121ef54dd4537ef59ae1fa7b6dcd0189f2"} Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.854840 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:56 crc kubenswrapper[4725]: I1202 13:20:56.854894 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:20:57 crc kubenswrapper[4725]: I1202 13:20:57.467823 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-bvlkl" podStartSLOduration=25.678436339 podStartE2EDuration="30.467802017s" podCreationTimestamp="2025-12-02 13:20:27 +0000 UTC" firstStartedPulling="2025-12-02 13:20:43.812617037 +0000 UTC m=+974.769258732" lastFinishedPulling="2025-12-02 13:20:48.601982715 +0000 UTC m=+979.558624410" observedRunningTime="2025-12-02 13:20:57.115368146 +0000 UTC m=+988.072009841" watchObservedRunningTime="2025-12-02 13:20:57.467802017 +0000 UTC m=+988.424443712" Dec 02 13:20:57 crc kubenswrapper[4725]: I1202 13:20:57.437875 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.97:5353: connect: connection refused" Dec 02 13:20:57 crc kubenswrapper[4725]: I1202 13:20:57.990602 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 02 13:20:58 crc kubenswrapper[4725]: I1202 13:20:58.115953 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.052432 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.247968 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.493240 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.558444 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-config\") pod \"0536fef7-fbd3-4633-a373-3c6e447d8a19\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.558592 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-dns-svc\") pod \"0536fef7-fbd3-4633-a373-3c6e447d8a19\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.558646 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlmcm\" (UniqueName: \"kubernetes.io/projected/0536fef7-fbd3-4633-a373-3c6e447d8a19-kube-api-access-zlmcm\") pod \"0536fef7-fbd3-4633-a373-3c6e447d8a19\" (UID: \"0536fef7-fbd3-4633-a373-3c6e447d8a19\") " Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.564481 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0536fef7-fbd3-4633-a373-3c6e447d8a19-kube-api-access-zlmcm" (OuterVolumeSpecName: "kube-api-access-zlmcm") pod "0536fef7-fbd3-4633-a373-3c6e447d8a19" (UID: "0536fef7-fbd3-4633-a373-3c6e447d8a19"). InnerVolumeSpecName "kube-api-access-zlmcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.600188 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-config" (OuterVolumeSpecName: "config") pod "0536fef7-fbd3-4633-a373-3c6e447d8a19" (UID: "0536fef7-fbd3-4633-a373-3c6e447d8a19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.608649 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0536fef7-fbd3-4633-a373-3c6e447d8a19" (UID: "0536fef7-fbd3-4633-a373-3c6e447d8a19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.663149 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.663187 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0536fef7-fbd3-4633-a373-3c6e447d8a19-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:00 crc kubenswrapper[4725]: I1202 13:21:00.663199 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlmcm\" (UniqueName: \"kubernetes.io/projected/0536fef7-fbd3-4633-a373-3c6e447d8a19-kube-api-access-zlmcm\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.003747 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.003882 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n78h67bhbchfh554h587h665h65bh9h698h5cfh5bdhcbh574hd6h594h5f6h546h65ch56bh575h5bbhcdh74h6ch7fh87h57dh685hd7hc9h55dq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-745xk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(51144a39-c6dd-4a5f-adfb-6d62cb459485): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.005586 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="51144a39-c6dd-4a5f-adfb-6d62cb459485" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.008505 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" event={"ID":"0536fef7-fbd3-4633-a373-3c6e447d8a19","Type":"ContainerDied","Data":"3e843a4ab2f8dd720336d5d8934cfaaf2f3f22aba440a0fc486199ccdc0baf2e"} Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.008566 4725 scope.go:117] "RemoveContainer" containerID="c4b8d491ed83ce2e962346bee5c24b02e7af38bb58e46c51daf0a8ca741b9f5e" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.008582 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qhxnm" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.008863 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.009021 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n678h69h5d7h544h6fhdfh85h556h579h68bhc9h645h557h5fch574h554h664h688hcbh597h5dchb9h5f8h5cch68hf9hf9hfch6bh9fh5d5hccq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9gpss,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(bad22ab8-0033-4a28-b7aa-65b1c285a989): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.010204 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="bad22ab8-0033-4a28-b7aa-65b1c285a989" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.029655 4725 scope.go:117] "RemoveContainer" containerID="936a3e684329aa1d63ab9776d68a76e8deaba48749057b4e2377a8bc56c0a6bc" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.046415 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhxnm"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.070960 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qhxnm"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.277176 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" path="/var/lib/kubelet/pods/0536fef7-fbd3-4633-a373-3c6e447d8a19/volumes" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.320899 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-81d8-account-create-update-tvqbp"] Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.321510 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="dnsmasq-dns" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.321577 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="dnsmasq-dns" Dec 02 13:21:01 crc kubenswrapper[4725]: E1202 13:21:01.321643 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="init" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.321701 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="init" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.321949 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="0536fef7-fbd3-4633-a373-3c6e447d8a19" containerName="dnsmasq-dns" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.322668 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.327549 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.330285 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-81d8-account-create-update-tvqbp"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.369020 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-f6d7c"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.370484 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.384674 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-f6d7c"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.475738 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-operator-scripts\") pod \"keystone-81d8-account-create-update-tvqbp\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.475852 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt7qm\" (UniqueName: \"kubernetes.io/projected/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-kube-api-access-lt7qm\") pod \"keystone-81d8-account-create-update-tvqbp\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.475912 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f6m4\" (UniqueName: \"kubernetes.io/projected/cf90bb11-ddb2-4433-82d3-b168999cc158-kube-api-access-8f6m4\") pod \"keystone-db-create-f6d7c\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.476170 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf90bb11-ddb2-4433-82d3-b168999cc158-operator-scripts\") pod \"keystone-db-create-f6d7c\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.577739 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt7qm\" (UniqueName: \"kubernetes.io/projected/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-kube-api-access-lt7qm\") pod \"keystone-81d8-account-create-update-tvqbp\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.577799 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f6m4\" (UniqueName: \"kubernetes.io/projected/cf90bb11-ddb2-4433-82d3-b168999cc158-kube-api-access-8f6m4\") pod \"keystone-db-create-f6d7c\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.577893 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf90bb11-ddb2-4433-82d3-b168999cc158-operator-scripts\") pod \"keystone-db-create-f6d7c\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.577921 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-operator-scripts\") pod \"keystone-81d8-account-create-update-tvqbp\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.578725 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf90bb11-ddb2-4433-82d3-b168999cc158-operator-scripts\") pod \"keystone-db-create-f6d7c\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.579009 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-operator-scripts\") pod \"keystone-81d8-account-create-update-tvqbp\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.599407 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f6m4\" (UniqueName: \"kubernetes.io/projected/cf90bb11-ddb2-4433-82d3-b168999cc158-kube-api-access-8f6m4\") pod \"keystone-db-create-f6d7c\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.602083 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt7qm\" (UniqueName: \"kubernetes.io/projected/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-kube-api-access-lt7qm\") pod \"keystone-81d8-account-create-update-tvqbp\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.641084 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-l9fs7"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.642336 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.642426 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.655614 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4596-account-create-update-nfdpv"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.656858 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.658244 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.665240 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-l9fs7"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.677795 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4596-account-create-update-nfdpv"] Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.689844 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.781424 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62f5f\" (UniqueName: \"kubernetes.io/projected/c55f5277-1f0c-450f-8d58-54d808005c47-kube-api-access-62f5f\") pod \"placement-db-create-l9fs7\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.781535 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c55f5277-1f0c-450f-8d58-54d808005c47-operator-scripts\") pod \"placement-db-create-l9fs7\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.781579 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793f192-d30b-47f0-a0e0-b31c07946211-operator-scripts\") pod \"placement-4596-account-create-update-nfdpv\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.781616 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rntb6\" (UniqueName: \"kubernetes.io/projected/9793f192-d30b-47f0-a0e0-b31c07946211-kube-api-access-rntb6\") pod \"placement-4596-account-create-update-nfdpv\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.883187 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62f5f\" (UniqueName: \"kubernetes.io/projected/c55f5277-1f0c-450f-8d58-54d808005c47-kube-api-access-62f5f\") pod \"placement-db-create-l9fs7\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.883287 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c55f5277-1f0c-450f-8d58-54d808005c47-operator-scripts\") pod \"placement-db-create-l9fs7\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.883332 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793f192-d30b-47f0-a0e0-b31c07946211-operator-scripts\") pod \"placement-4596-account-create-update-nfdpv\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.883378 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rntb6\" (UniqueName: \"kubernetes.io/projected/9793f192-d30b-47f0-a0e0-b31c07946211-kube-api-access-rntb6\") pod \"placement-4596-account-create-update-nfdpv\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.884295 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c55f5277-1f0c-450f-8d58-54d808005c47-operator-scripts\") pod \"placement-db-create-l9fs7\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.884812 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793f192-d30b-47f0-a0e0-b31c07946211-operator-scripts\") pod \"placement-4596-account-create-update-nfdpv\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.900396 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rntb6\" (UniqueName: \"kubernetes.io/projected/9793f192-d30b-47f0-a0e0-b31c07946211-kube-api-access-rntb6\") pod \"placement-4596-account-create-update-nfdpv\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:01 crc kubenswrapper[4725]: I1202 13:21:01.903133 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62f5f\" (UniqueName: \"kubernetes.io/projected/c55f5277-1f0c-450f-8d58-54d808005c47-kube-api-access-62f5f\") pod \"placement-db-create-l9fs7\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:02 crc kubenswrapper[4725]: I1202 13:21:02.101999 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:02 crc kubenswrapper[4725]: I1202 13:21:02.140315 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:02 crc kubenswrapper[4725]: E1202 13:21:02.171606 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bad22ab8-0033-4a28-b7aa-65b1c285a989" Dec 02 13:21:02 crc kubenswrapper[4725]: E1202 13:21:02.172898 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="51144a39-c6dd-4a5f-adfb-6d62cb459485" Dec 02 13:21:02 crc kubenswrapper[4725]: I1202 13:21:02.362747 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-81d8-account-create-update-tvqbp"] Dec 02 13:21:02 crc kubenswrapper[4725]: I1202 13:21:02.823125 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-f6d7c"] Dec 02 13:21:02 crc kubenswrapper[4725]: W1202 13:21:02.836126 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf90bb11_ddb2_4433_82d3_b168999cc158.slice/crio-bcf6ff3a0353a51fa923029be319022e0b66dca5a87f167089bc3fd39abe34c1 WatchSource:0}: Error finding container bcf6ff3a0353a51fa923029be319022e0b66dca5a87f167089bc3fd39abe34c1: Status 404 returned error can't find the container with id bcf6ff3a0353a51fa923029be319022e0b66dca5a87f167089bc3fd39abe34c1 Dec 02 13:21:02 crc kubenswrapper[4725]: I1202 13:21:02.942091 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-l9fs7"] Dec 02 13:21:02 crc kubenswrapper[4725]: W1202 13:21:02.943731 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc55f5277_1f0c_450f_8d58_54d808005c47.slice/crio-161e3f444e81119d90be42729a0d60395b807fc02c5a65f739e929b94432dc53 WatchSource:0}: Error finding container 161e3f444e81119d90be42729a0d60395b807fc02c5a65f739e929b94432dc53: Status 404 returned error can't find the container with id 161e3f444e81119d90be42729a0d60395b807fc02c5a65f739e929b94432dc53 Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.185444 4725 generic.go:334] "Generic (PLEG): container finished" podID="90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" containerID="576ae5c7c4c6166ce9e548c491b67e056fc5d7ca1b77bee37ff5c5b83cfe8789" exitCode=0 Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.185517 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81d8-account-create-update-tvqbp" event={"ID":"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5","Type":"ContainerDied","Data":"576ae5c7c4c6166ce9e548c491b67e056fc5d7ca1b77bee37ff5c5b83cfe8789"} Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.185581 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81d8-account-create-update-tvqbp" event={"ID":"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5","Type":"ContainerStarted","Data":"7e4042d4844b121af6d13e811907939e940ee0818122b6a7da4a33146960f0e9"} Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.188412 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l9fs7" event={"ID":"c55f5277-1f0c-450f-8d58-54d808005c47","Type":"ContainerStarted","Data":"835ac8643da173853954018e41884fc776adee8dc9425aeb33fda07cf9f77ddf"} Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.188444 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l9fs7" event={"ID":"c55f5277-1f0c-450f-8d58-54d808005c47","Type":"ContainerStarted","Data":"161e3f444e81119d90be42729a0d60395b807fc02c5a65f739e929b94432dc53"} Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.190381 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f6d7c" event={"ID":"cf90bb11-ddb2-4433-82d3-b168999cc158","Type":"ContainerStarted","Data":"82af7c2f7a6a510d4d8c7d20856cc39062a65d8de1b75c7144e34c99c62b62f8"} Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.190421 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f6d7c" event={"ID":"cf90bb11-ddb2-4433-82d3-b168999cc158","Type":"ContainerStarted","Data":"bcf6ff3a0353a51fa923029be319022e0b66dca5a87f167089bc3fd39abe34c1"} Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.218587 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-l9fs7" podStartSLOduration=2.218569259 podStartE2EDuration="2.218569259s" podCreationTimestamp="2025-12-02 13:21:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:03.218374384 +0000 UTC m=+994.175016079" watchObservedRunningTime="2025-12-02 13:21:03.218569259 +0000 UTC m=+994.175210954" Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.269216 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-f6d7c" podStartSLOduration=2.269197346 podStartE2EDuration="2.269197346s" podCreationTimestamp="2025-12-02 13:21:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:03.266411826 +0000 UTC m=+994.223053521" watchObservedRunningTime="2025-12-02 13:21:03.269197346 +0000 UTC m=+994.225839041" Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.309395 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4596-account-create-update-nfdpv"] Dec 02 13:21:03 crc kubenswrapper[4725]: I1202 13:21:03.682931 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 02 13:21:03 crc kubenswrapper[4725]: E1202 13:21:03.684612 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bad22ab8-0033-4a28-b7aa-65b1c285a989" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.191115 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxp5n"] Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.193826 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.210104 4725 generic.go:334] "Generic (PLEG): container finished" podID="cf90bb11-ddb2-4433-82d3-b168999cc158" containerID="82af7c2f7a6a510d4d8c7d20856cc39062a65d8de1b75c7144e34c99c62b62f8" exitCode=0 Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.210200 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f6d7c" event={"ID":"cf90bb11-ddb2-4433-82d3-b168999cc158","Type":"ContainerDied","Data":"82af7c2f7a6a510d4d8c7d20856cc39062a65d8de1b75c7144e34c99c62b62f8"} Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.213715 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxp5n"] Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.222424 4725 generic.go:334] "Generic (PLEG): container finished" podID="9793f192-d30b-47f0-a0e0-b31c07946211" containerID="d98e99acffb691b6d57cf4fadb743c9f0b12e32adc9fb5dd0dc66117bfb3e57b" exitCode=0 Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.222503 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4596-account-create-update-nfdpv" event={"ID":"9793f192-d30b-47f0-a0e0-b31c07946211","Type":"ContainerDied","Data":"d98e99acffb691b6d57cf4fadb743c9f0b12e32adc9fb5dd0dc66117bfb3e57b"} Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.222529 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4596-account-create-update-nfdpv" event={"ID":"9793f192-d30b-47f0-a0e0-b31c07946211","Type":"ContainerStarted","Data":"9115395aec2115fbf0003f455dc430ed82a417a1c4269828a365220cf02c2b85"} Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.230780 4725 generic.go:334] "Generic (PLEG): container finished" podID="c55f5277-1f0c-450f-8d58-54d808005c47" containerID="835ac8643da173853954018e41884fc776adee8dc9425aeb33fda07cf9f77ddf" exitCode=0 Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.231110 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l9fs7" event={"ID":"c55f5277-1f0c-450f-8d58-54d808005c47","Type":"ContainerDied","Data":"835ac8643da173853954018e41884fc776adee8dc9425aeb33fda07cf9f77ddf"} Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.285511 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cdrl\" (UniqueName: \"kubernetes.io/projected/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-kube-api-access-7cdrl\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.285587 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.285618 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-config\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.387671 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cdrl\" (UniqueName: \"kubernetes.io/projected/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-kube-api-access-7cdrl\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.387726 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.387770 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-config\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.391684 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-config\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.392071 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.414001 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cdrl\" (UniqueName: \"kubernetes.io/projected/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-kube-api-access-7cdrl\") pod \"dnsmasq-dns-7cb5889db5-fxp5n\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.533659 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.638418 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.682141 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 02 13:21:04 crc kubenswrapper[4725]: E1202 13:21:04.689902 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bad22ab8-0033-4a28-b7aa-65b1c285a989" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.749975 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.792760 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt7qm\" (UniqueName: \"kubernetes.io/projected/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-kube-api-access-lt7qm\") pod \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.792796 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-operator-scripts\") pod \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\" (UID: \"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5\") " Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.793712 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" (UID: "90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.796854 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-kube-api-access-lt7qm" (OuterVolumeSpecName: "kube-api-access-lt7qm") pod "90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" (UID: "90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5"). InnerVolumeSpecName "kube-api-access-lt7qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.816065 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 02 13:21:04 crc kubenswrapper[4725]: E1202 13:21:04.817819 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="51144a39-c6dd-4a5f-adfb-6d62cb459485" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.860752 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.894611 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt7qm\" (UniqueName: \"kubernetes.io/projected/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-kube-api-access-lt7qm\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.894673 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:04 crc kubenswrapper[4725]: I1202 13:21:04.980968 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxp5n"] Dec 02 13:21:04 crc kubenswrapper[4725]: W1202 13:21:04.985699 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a0325e3_d5eb_4a03_aa3a_bbc172697b9b.slice/crio-6d5c5d52eb2247b600ffb5fa113b7bc2ddb25d5ffb0b01f616c7620f8b4fdad1 WatchSource:0}: Error finding container 6d5c5d52eb2247b600ffb5fa113b7bc2ddb25d5ffb0b01f616c7620f8b4fdad1: Status 404 returned error can't find the container with id 6d5c5d52eb2247b600ffb5fa113b7bc2ddb25d5ffb0b01f616c7620f8b4fdad1 Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.241288 4725 generic.go:334] "Generic (PLEG): container finished" podID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerID="226de640a1f4a608f4933ef43c8ea968e0a3d8e31ac183e3b803a7cd41ea4967" exitCode=0 Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.241407 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" event={"ID":"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b","Type":"ContainerDied","Data":"226de640a1f4a608f4933ef43c8ea968e0a3d8e31ac183e3b803a7cd41ea4967"} Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.241622 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" event={"ID":"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b","Type":"ContainerStarted","Data":"6d5c5d52eb2247b600ffb5fa113b7bc2ddb25d5ffb0b01f616c7620f8b4fdad1"} Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.245523 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-81d8-account-create-update-tvqbp" event={"ID":"90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5","Type":"ContainerDied","Data":"7e4042d4844b121af6d13e811907939e940ee0818122b6a7da4a33146960f0e9"} Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.245551 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e4042d4844b121af6d13e811907939e940ee0818122b6a7da4a33146960f0e9" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.246021 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.246758 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 02 13:21:05 crc kubenswrapper[4725]: E1202 13:21:05.250472 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="51144a39-c6dd-4a5f-adfb-6d62cb459485" Dec 02 13:21:05 crc kubenswrapper[4725]: E1202 13:21:05.250514 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bad22ab8-0033-4a28-b7aa-65b1c285a989" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.360577 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.403109 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.427844 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 02 13:21:05 crc kubenswrapper[4725]: E1202 13:21:05.428327 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" containerName="mariadb-account-create-update" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.428344 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" containerName="mariadb-account-create-update" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.428578 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" containerName="mariadb-account-create-update" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.437321 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.441400 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.441488 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-dg7v9" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.441665 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.441717 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.462832 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.519545 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-cache\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.519599 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m4zk\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-kube-api-access-2m4zk\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.519651 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.519692 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.520058 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-lock\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.629767 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.629827 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: E1202 13:21:05.630061 4725 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 13:21:05 crc kubenswrapper[4725]: E1202 13:21:05.630092 4725 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 13:21:05 crc kubenswrapper[4725]: E1202 13:21:05.630144 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift podName:c3fe5753-c7c5-450a-b4e6-0065deae5cd9 nodeName:}" failed. No retries permitted until 2025-12-02 13:21:06.130127883 +0000 UTC m=+997.086769578 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift") pod "swift-storage-0" (UID: "c3fe5753-c7c5-450a-b4e6-0065deae5cd9") : configmap "swift-ring-files" not found Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.630535 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.633909 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-lock\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.634301 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-cache\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.634372 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m4zk\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-kube-api-access-2m4zk\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.635206 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-cache\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.635407 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-lock\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.659319 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m4zk\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-kube-api-access-2m4zk\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.676838 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.810882 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.813794 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.826326 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.939910 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f6m4\" (UniqueName: \"kubernetes.io/projected/cf90bb11-ddb2-4433-82d3-b168999cc158-kube-api-access-8f6m4\") pod \"cf90bb11-ddb2-4433-82d3-b168999cc158\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.940196 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c55f5277-1f0c-450f-8d58-54d808005c47-operator-scripts\") pod \"c55f5277-1f0c-450f-8d58-54d808005c47\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.940333 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rntb6\" (UniqueName: \"kubernetes.io/projected/9793f192-d30b-47f0-a0e0-b31c07946211-kube-api-access-rntb6\") pod \"9793f192-d30b-47f0-a0e0-b31c07946211\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.940373 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf90bb11-ddb2-4433-82d3-b168999cc158-operator-scripts\") pod \"cf90bb11-ddb2-4433-82d3-b168999cc158\" (UID: \"cf90bb11-ddb2-4433-82d3-b168999cc158\") " Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.940413 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793f192-d30b-47f0-a0e0-b31c07946211-operator-scripts\") pod \"9793f192-d30b-47f0-a0e0-b31c07946211\" (UID: \"9793f192-d30b-47f0-a0e0-b31c07946211\") " Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.940441 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62f5f\" (UniqueName: \"kubernetes.io/projected/c55f5277-1f0c-450f-8d58-54d808005c47-kube-api-access-62f5f\") pod \"c55f5277-1f0c-450f-8d58-54d808005c47\" (UID: \"c55f5277-1f0c-450f-8d58-54d808005c47\") " Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.941443 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf90bb11-ddb2-4433-82d3-b168999cc158-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf90bb11-ddb2-4433-82d3-b168999cc158" (UID: "cf90bb11-ddb2-4433-82d3-b168999cc158"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.941433 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c55f5277-1f0c-450f-8d58-54d808005c47-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c55f5277-1f0c-450f-8d58-54d808005c47" (UID: "c55f5277-1f0c-450f-8d58-54d808005c47"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.941725 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9793f192-d30b-47f0-a0e0-b31c07946211-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9793f192-d30b-47f0-a0e0-b31c07946211" (UID: "9793f192-d30b-47f0-a0e0-b31c07946211"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.946247 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c55f5277-1f0c-450f-8d58-54d808005c47-kube-api-access-62f5f" (OuterVolumeSpecName: "kube-api-access-62f5f") pod "c55f5277-1f0c-450f-8d58-54d808005c47" (UID: "c55f5277-1f0c-450f-8d58-54d808005c47"). InnerVolumeSpecName "kube-api-access-62f5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.946610 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf90bb11-ddb2-4433-82d3-b168999cc158-kube-api-access-8f6m4" (OuterVolumeSpecName: "kube-api-access-8f6m4") pod "cf90bb11-ddb2-4433-82d3-b168999cc158" (UID: "cf90bb11-ddb2-4433-82d3-b168999cc158"). InnerVolumeSpecName "kube-api-access-8f6m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:05 crc kubenswrapper[4725]: I1202 13:21:05.948027 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9793f192-d30b-47f0-a0e0-b31c07946211-kube-api-access-rntb6" (OuterVolumeSpecName: "kube-api-access-rntb6") pod "9793f192-d30b-47f0-a0e0-b31c07946211" (UID: "9793f192-d30b-47f0-a0e0-b31c07946211"). InnerVolumeSpecName "kube-api-access-rntb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.042773 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c55f5277-1f0c-450f-8d58-54d808005c47-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.042820 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rntb6\" (UniqueName: \"kubernetes.io/projected/9793f192-d30b-47f0-a0e0-b31c07946211-kube-api-access-rntb6\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.042838 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf90bb11-ddb2-4433-82d3-b168999cc158-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.042854 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9793f192-d30b-47f0-a0e0-b31c07946211-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.042866 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62f5f\" (UniqueName: \"kubernetes.io/projected/c55f5277-1f0c-450f-8d58-54d808005c47-kube-api-access-62f5f\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.042878 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f6m4\" (UniqueName: \"kubernetes.io/projected/cf90bb11-ddb2-4433-82d3-b168999cc158-kube-api-access-8f6m4\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.143898 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:06 crc kubenswrapper[4725]: E1202 13:21:06.144170 4725 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 13:21:06 crc kubenswrapper[4725]: E1202 13:21:06.144190 4725 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 13:21:06 crc kubenswrapper[4725]: E1202 13:21:06.144244 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift podName:c3fe5753-c7c5-450a-b4e6-0065deae5cd9 nodeName:}" failed. No retries permitted until 2025-12-02 13:21:07.144229617 +0000 UTC m=+998.100871312 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift") pod "swift-storage-0" (UID: "c3fe5753-c7c5-450a-b4e6-0065deae5cd9") : configmap "swift-ring-files" not found Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.254668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4596-account-create-update-nfdpv" event={"ID":"9793f192-d30b-47f0-a0e0-b31c07946211","Type":"ContainerDied","Data":"9115395aec2115fbf0003f455dc430ed82a417a1c4269828a365220cf02c2b85"} Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.254693 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4596-account-create-update-nfdpv" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.254773 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9115395aec2115fbf0003f455dc430ed82a417a1c4269828a365220cf02c2b85" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.256403 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-l9fs7" event={"ID":"c55f5277-1f0c-450f-8d58-54d808005c47","Type":"ContainerDied","Data":"161e3f444e81119d90be42729a0d60395b807fc02c5a65f739e929b94432dc53"} Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.256441 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="161e3f444e81119d90be42729a0d60395b807fc02c5a65f739e929b94432dc53" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.256534 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-l9fs7" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.259883 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-f6d7c" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.271337 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-f6d7c" event={"ID":"cf90bb11-ddb2-4433-82d3-b168999cc158","Type":"ContainerDied","Data":"bcf6ff3a0353a51fa923029be319022e0b66dca5a87f167089bc3fd39abe34c1"} Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.271372 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcf6ff3a0353a51fa923029be319022e0b66dca5a87f167089bc3fd39abe34c1" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.282028 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" event={"ID":"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b","Type":"ContainerStarted","Data":"164e78819f34d4adc0d2fb9200dadd42794af848fca8c10e95ea1c7b0f5abf18"} Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.282427 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:06 crc kubenswrapper[4725]: E1202 13:21:06.286038 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="51144a39-c6dd-4a5f-adfb-6d62cb459485" Dec 02 13:21:06 crc kubenswrapper[4725]: E1202 13:21:06.290934 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bad22ab8-0033-4a28-b7aa-65b1c285a989" Dec 02 13:21:06 crc kubenswrapper[4725]: I1202 13:21:06.316206 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" podStartSLOduration=2.316178577 podStartE2EDuration="2.316178577s" podCreationTimestamp="2025-12-02 13:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:06.30584356 +0000 UTC m=+997.262485255" watchObservedRunningTime="2025-12-02 13:21:06.316178577 +0000 UTC m=+997.272820272" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.159401 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.159601 4725 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.159874 4725 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.159941 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift podName:c3fe5753-c7c5-450a-b4e6-0065deae5cd9 nodeName:}" failed. No retries permitted until 2025-12-02 13:21:09.159921905 +0000 UTC m=+1000.116563600 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift") pod "swift-storage-0" (UID: "c3fe5753-c7c5-450a-b4e6-0065deae5cd9") : configmap "swift-ring-files" not found Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.289532 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="51144a39-c6dd-4a5f-adfb-6d62cb459485" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.489987 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-vd2kp"] Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.490318 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf90bb11-ddb2-4433-82d3-b168999cc158" containerName="mariadb-database-create" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.490334 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf90bb11-ddb2-4433-82d3-b168999cc158" containerName="mariadb-database-create" Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.490363 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c55f5277-1f0c-450f-8d58-54d808005c47" containerName="mariadb-database-create" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.490371 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c55f5277-1f0c-450f-8d58-54d808005c47" containerName="mariadb-database-create" Dec 02 13:21:07 crc kubenswrapper[4725]: E1202 13:21:07.490380 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9793f192-d30b-47f0-a0e0-b31c07946211" containerName="mariadb-account-create-update" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.490387 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9793f192-d30b-47f0-a0e0-b31c07946211" containerName="mariadb-account-create-update" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.490556 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf90bb11-ddb2-4433-82d3-b168999cc158" containerName="mariadb-database-create" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.490570 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="c55f5277-1f0c-450f-8d58-54d808005c47" containerName="mariadb-database-create" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.490583 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="9793f192-d30b-47f0-a0e0-b31c07946211" containerName="mariadb-account-create-update" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.491113 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.509979 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-vd2kp"] Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.566767 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zpnm\" (UniqueName: \"kubernetes.io/projected/f145bf96-9032-4d92-8aac-9ebe20153fca-kube-api-access-2zpnm\") pod \"glance-db-create-vd2kp\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.566890 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f145bf96-9032-4d92-8aac-9ebe20153fca-operator-scripts\") pod \"glance-db-create-vd2kp\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.613882 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a1c9-account-create-update-jzlj6"] Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.614971 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.620231 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.627758 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a1c9-account-create-update-jzlj6"] Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.668415 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zpnm\" (UniqueName: \"kubernetes.io/projected/f145bf96-9032-4d92-8aac-9ebe20153fca-kube-api-access-2zpnm\") pod \"glance-db-create-vd2kp\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.668516 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-operator-scripts\") pod \"glance-a1c9-account-create-update-jzlj6\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.668540 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f145bf96-9032-4d92-8aac-9ebe20153fca-operator-scripts\") pod \"glance-db-create-vd2kp\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.668643 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmsfz\" (UniqueName: \"kubernetes.io/projected/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-kube-api-access-zmsfz\") pod \"glance-a1c9-account-create-update-jzlj6\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.669303 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f145bf96-9032-4d92-8aac-9ebe20153fca-operator-scripts\") pod \"glance-db-create-vd2kp\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.686864 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zpnm\" (UniqueName: \"kubernetes.io/projected/f145bf96-9032-4d92-8aac-9ebe20153fca-kube-api-access-2zpnm\") pod \"glance-db-create-vd2kp\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.770510 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-operator-scripts\") pod \"glance-a1c9-account-create-update-jzlj6\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.770998 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmsfz\" (UniqueName: \"kubernetes.io/projected/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-kube-api-access-zmsfz\") pod \"glance-a1c9-account-create-update-jzlj6\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.771412 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-operator-scripts\") pod \"glance-a1c9-account-create-update-jzlj6\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.789318 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmsfz\" (UniqueName: \"kubernetes.io/projected/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-kube-api-access-zmsfz\") pod \"glance-a1c9-account-create-update-jzlj6\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.808373 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:07 crc kubenswrapper[4725]: I1202 13:21:07.929952 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:08 crc kubenswrapper[4725]: I1202 13:21:08.265247 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-vd2kp"] Dec 02 13:21:08 crc kubenswrapper[4725]: I1202 13:21:08.294327 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-vd2kp" event={"ID":"f145bf96-9032-4d92-8aac-9ebe20153fca","Type":"ContainerStarted","Data":"d6b1c332e771fa26f0df732041377129e0e13f60326532919f59c56dee42ec54"} Dec 02 13:21:08 crc kubenswrapper[4725]: W1202 13:21:08.406676 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16ffbb6e_2974_43d3_aef7_6ff8b1a62aab.slice/crio-cd1ff65ab5975cd19b833e5846c59a02f0b945229e7ab9fec78c0eb35b61320f WatchSource:0}: Error finding container cd1ff65ab5975cd19b833e5846c59a02f0b945229e7ab9fec78c0eb35b61320f: Status 404 returned error can't find the container with id cd1ff65ab5975cd19b833e5846c59a02f0b945229e7ab9fec78c0eb35b61320f Dec 02 13:21:08 crc kubenswrapper[4725]: I1202 13:21:08.415893 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a1c9-account-create-update-jzlj6"] Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.193603 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:09 crc kubenswrapper[4725]: E1202 13:21:09.193876 4725 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 13:21:09 crc kubenswrapper[4725]: E1202 13:21:09.194008 4725 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 13:21:09 crc kubenswrapper[4725]: E1202 13:21:09.194083 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift podName:c3fe5753-c7c5-450a-b4e6-0065deae5cd9 nodeName:}" failed. No retries permitted until 2025-12-02 13:21:13.194063239 +0000 UTC m=+1004.150704934 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift") pod "swift-storage-0" (UID: "c3fe5753-c7c5-450a-b4e6-0065deae5cd9") : configmap "swift-ring-files" not found Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.307219 4725 generic.go:334] "Generic (PLEG): container finished" podID="f145bf96-9032-4d92-8aac-9ebe20153fca" containerID="940a93f04e76a7ec1949b2d4e141548dc960bf004222efffd4d2475fc5a64529" exitCode=0 Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.307305 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-vd2kp" event={"ID":"f145bf96-9032-4d92-8aac-9ebe20153fca","Type":"ContainerDied","Data":"940a93f04e76a7ec1949b2d4e141548dc960bf004222efffd4d2475fc5a64529"} Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.308641 4725 generic.go:334] "Generic (PLEG): container finished" podID="16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" containerID="18e29ee21f03cbf03585031abc2d12a9b72e9ccc532e1d82569a0c8dbba73d37" exitCode=0 Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.308679 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a1c9-account-create-update-jzlj6" event={"ID":"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab","Type":"ContainerDied","Data":"18e29ee21f03cbf03585031abc2d12a9b72e9ccc532e1d82569a0c8dbba73d37"} Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.308709 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a1c9-account-create-update-jzlj6" event={"ID":"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab","Type":"ContainerStarted","Data":"cd1ff65ab5975cd19b833e5846c59a02f0b945229e7ab9fec78c0eb35b61320f"} Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.340168 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-z72np"] Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.341374 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.344928 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.347717 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.350040 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.359690 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-z72np"] Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403410 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-dispersionconf\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403479 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-combined-ca-bundle\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403545 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-swiftconf\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403581 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbbdh\" (UniqueName: \"kubernetes.io/projected/dda4b826-1871-43b5-808d-45d84f2f2483-kube-api-access-cbbdh\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403599 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-ring-data-devices\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403641 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-scripts\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.403658 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dda4b826-1871-43b5-808d-45d84f2f2483-etc-swift\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505328 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-swiftconf\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505437 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbbdh\" (UniqueName: \"kubernetes.io/projected/dda4b826-1871-43b5-808d-45d84f2f2483-kube-api-access-cbbdh\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505483 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-ring-data-devices\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505532 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-scripts\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505561 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dda4b826-1871-43b5-808d-45d84f2f2483-etc-swift\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505650 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-dispersionconf\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.505701 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-combined-ca-bundle\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.506159 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dda4b826-1871-43b5-808d-45d84f2f2483-etc-swift\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.506420 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-ring-data-devices\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.506449 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-scripts\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.511825 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-combined-ca-bundle\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.519952 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-swiftconf\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.523099 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbbdh\" (UniqueName: \"kubernetes.io/projected/dda4b826-1871-43b5-808d-45d84f2f2483-kube-api-access-cbbdh\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.523791 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-dispersionconf\") pod \"swift-ring-rebalance-z72np\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:09 crc kubenswrapper[4725]: I1202 13:21:09.657478 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.072126 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-z72np"] Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.316529 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z72np" event={"ID":"dda4b826-1871-43b5-808d-45d84f2f2483","Type":"ContainerStarted","Data":"55b6b19ddd9dc165b7cef63170558a4998fce79bd4f646f2ecf30bebbe5d08c3"} Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.637571 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.729407 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-operator-scripts\") pod \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.729494 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmsfz\" (UniqueName: \"kubernetes.io/projected/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-kube-api-access-zmsfz\") pod \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\" (UID: \"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab\") " Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.732115 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" (UID: "16ffbb6e-2974-43d3-aef7-6ff8b1a62aab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.734724 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.739162 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-kube-api-access-zmsfz" (OuterVolumeSpecName: "kube-api-access-zmsfz") pod "16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" (UID: "16ffbb6e-2974-43d3-aef7-6ff8b1a62aab"). InnerVolumeSpecName "kube-api-access-zmsfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.831173 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f145bf96-9032-4d92-8aac-9ebe20153fca-operator-scripts\") pod \"f145bf96-9032-4d92-8aac-9ebe20153fca\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.831333 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zpnm\" (UniqueName: \"kubernetes.io/projected/f145bf96-9032-4d92-8aac-9ebe20153fca-kube-api-access-2zpnm\") pod \"f145bf96-9032-4d92-8aac-9ebe20153fca\" (UID: \"f145bf96-9032-4d92-8aac-9ebe20153fca\") " Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.831729 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f145bf96-9032-4d92-8aac-9ebe20153fca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f145bf96-9032-4d92-8aac-9ebe20153fca" (UID: "f145bf96-9032-4d92-8aac-9ebe20153fca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.831931 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmsfz\" (UniqueName: \"kubernetes.io/projected/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-kube-api-access-zmsfz\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.831945 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f145bf96-9032-4d92-8aac-9ebe20153fca-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.831955 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.834245 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f145bf96-9032-4d92-8aac-9ebe20153fca-kube-api-access-2zpnm" (OuterVolumeSpecName: "kube-api-access-2zpnm") pod "f145bf96-9032-4d92-8aac-9ebe20153fca" (UID: "f145bf96-9032-4d92-8aac-9ebe20153fca"). InnerVolumeSpecName "kube-api-access-2zpnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:10 crc kubenswrapper[4725]: I1202 13:21:10.933913 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zpnm\" (UniqueName: \"kubernetes.io/projected/f145bf96-9032-4d92-8aac-9ebe20153fca-kube-api-access-2zpnm\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:11 crc kubenswrapper[4725]: I1202 13:21:11.334074 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a1c9-account-create-update-jzlj6" event={"ID":"16ffbb6e-2974-43d3-aef7-6ff8b1a62aab","Type":"ContainerDied","Data":"cd1ff65ab5975cd19b833e5846c59a02f0b945229e7ab9fec78c0eb35b61320f"} Dec 02 13:21:11 crc kubenswrapper[4725]: I1202 13:21:11.334444 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd1ff65ab5975cd19b833e5846c59a02f0b945229e7ab9fec78c0eb35b61320f" Dec 02 13:21:11 crc kubenswrapper[4725]: I1202 13:21:11.334451 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a1c9-account-create-update-jzlj6" Dec 02 13:21:11 crc kubenswrapper[4725]: I1202 13:21:11.337436 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-vd2kp" event={"ID":"f145bf96-9032-4d92-8aac-9ebe20153fca","Type":"ContainerDied","Data":"d6b1c332e771fa26f0df732041377129e0e13f60326532919f59c56dee42ec54"} Dec 02 13:21:11 crc kubenswrapper[4725]: I1202 13:21:11.337610 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-vd2kp" Dec 02 13:21:11 crc kubenswrapper[4725]: I1202 13:21:11.338517 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6b1c332e771fa26f0df732041377129e0e13f60326532919f59c56dee42ec54" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.728772 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-7d27g"] Dec 02 13:21:12 crc kubenswrapper[4725]: E1202 13:21:12.731748 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" containerName="mariadb-account-create-update" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.731862 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" containerName="mariadb-account-create-update" Dec 02 13:21:12 crc kubenswrapper[4725]: E1202 13:21:12.731962 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f145bf96-9032-4d92-8aac-9ebe20153fca" containerName="mariadb-database-create" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.732040 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f145bf96-9032-4d92-8aac-9ebe20153fca" containerName="mariadb-database-create" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.732364 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f145bf96-9032-4d92-8aac-9ebe20153fca" containerName="mariadb-database-create" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.732502 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" containerName="mariadb-account-create-update" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.735069 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.742514 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.743852 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-c5cmk" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.746385 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7d27g"] Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.765198 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lsp9\" (UniqueName: \"kubernetes.io/projected/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-kube-api-access-6lsp9\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.765319 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-combined-ca-bundle\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.765384 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-db-sync-config-data\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.765445 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-config-data\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.867536 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-db-sync-config-data\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.867641 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-config-data\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.867688 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lsp9\" (UniqueName: \"kubernetes.io/projected/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-kube-api-access-6lsp9\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.867746 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-combined-ca-bundle\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.877860 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-config-data\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.881077 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-db-sync-config-data\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.889755 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-combined-ca-bundle\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:12 crc kubenswrapper[4725]: I1202 13:21:12.896173 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lsp9\" (UniqueName: \"kubernetes.io/projected/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-kube-api-access-6lsp9\") pod \"glance-db-sync-7d27g\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:13 crc kubenswrapper[4725]: I1202 13:21:13.061428 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7d27g" Dec 02 13:21:13 crc kubenswrapper[4725]: I1202 13:21:13.273508 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:13 crc kubenswrapper[4725]: E1202 13:21:13.273779 4725 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 13:21:13 crc kubenswrapper[4725]: E1202 13:21:13.273801 4725 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 13:21:13 crc kubenswrapper[4725]: E1202 13:21:13.273864 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift podName:c3fe5753-c7c5-450a-b4e6-0065deae5cd9 nodeName:}" failed. No retries permitted until 2025-12-02 13:21:21.273845454 +0000 UTC m=+1012.230487149 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift") pod "swift-storage-0" (UID: "c3fe5753-c7c5-450a-b4e6-0065deae5cd9") : configmap "swift-ring-files" not found Dec 02 13:21:14 crc kubenswrapper[4725]: I1202 13:21:14.535561 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:14 crc kubenswrapper[4725]: I1202 13:21:14.617126 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-s7cw4"] Dec 02 13:21:14 crc kubenswrapper[4725]: I1202 13:21:14.617663 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" podUID="08951dcf-7810-40b8-be36-9003fe20022b" containerName="dnsmasq-dns" containerID="cri-o://100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df" gracePeriod=10 Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.366215 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.378218 4725 generic.go:334] "Generic (PLEG): container finished" podID="08951dcf-7810-40b8-be36-9003fe20022b" containerID="100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df" exitCode=0 Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.378274 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" event={"ID":"08951dcf-7810-40b8-be36-9003fe20022b","Type":"ContainerDied","Data":"100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df"} Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.378313 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" event={"ID":"08951dcf-7810-40b8-be36-9003fe20022b","Type":"ContainerDied","Data":"811189f07385b84b3ebff449dd93e8dbf38172cf04d1936489802c288c4adfb7"} Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.378338 4725 scope.go:117] "RemoveContainer" containerID="100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.378551 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-s7cw4" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.408052 4725 scope.go:117] "RemoveContainer" containerID="73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.408085 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-config\") pod \"08951dcf-7810-40b8-be36-9003fe20022b\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.408735 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rlj4\" (UniqueName: \"kubernetes.io/projected/08951dcf-7810-40b8-be36-9003fe20022b-kube-api-access-7rlj4\") pod \"08951dcf-7810-40b8-be36-9003fe20022b\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.408928 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-dns-svc\") pod \"08951dcf-7810-40b8-be36-9003fe20022b\" (UID: \"08951dcf-7810-40b8-be36-9003fe20022b\") " Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.413084 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08951dcf-7810-40b8-be36-9003fe20022b-kube-api-access-7rlj4" (OuterVolumeSpecName: "kube-api-access-7rlj4") pod "08951dcf-7810-40b8-be36-9003fe20022b" (UID: "08951dcf-7810-40b8-be36-9003fe20022b"). InnerVolumeSpecName "kube-api-access-7rlj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.432478 4725 scope.go:117] "RemoveContainer" containerID="100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df" Dec 02 13:21:15 crc kubenswrapper[4725]: E1202 13:21:15.432936 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df\": container with ID starting with 100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df not found: ID does not exist" containerID="100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.432976 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df"} err="failed to get container status \"100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df\": rpc error: code = NotFound desc = could not find container \"100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df\": container with ID starting with 100ceec1c2e5d68124a711f718969d10bf9b8f5a670e3b542828af0204c016df not found: ID does not exist" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.432995 4725 scope.go:117] "RemoveContainer" containerID="73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce" Dec 02 13:21:15 crc kubenswrapper[4725]: E1202 13:21:15.433282 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce\": container with ID starting with 73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce not found: ID does not exist" containerID="73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.433310 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce"} err="failed to get container status \"73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce\": rpc error: code = NotFound desc = could not find container \"73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce\": container with ID starting with 73f1b386ff7f64b7b605b2e4cdb924543d514a974740678f478e4bd6569c04ce not found: ID does not exist" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.446924 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08951dcf-7810-40b8-be36-9003fe20022b" (UID: "08951dcf-7810-40b8-be36-9003fe20022b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.451949 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-config" (OuterVolumeSpecName: "config") pod "08951dcf-7810-40b8-be36-9003fe20022b" (UID: "08951dcf-7810-40b8-be36-9003fe20022b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.511392 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.511433 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08951dcf-7810-40b8-be36-9003fe20022b-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.511445 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rlj4\" (UniqueName: \"kubernetes.io/projected/08951dcf-7810-40b8-be36-9003fe20022b-kube-api-access-7rlj4\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.718998 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7d27g"] Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.726745 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-s7cw4"] Dec 02 13:21:15 crc kubenswrapper[4725]: I1202 13:21:15.733687 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-s7cw4"] Dec 02 13:21:16 crc kubenswrapper[4725]: I1202 13:21:16.387826 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z72np" event={"ID":"dda4b826-1871-43b5-808d-45d84f2f2483","Type":"ContainerStarted","Data":"41ee4eaa3ba651f62974272941794fd14a4c40b5255439aee5823eebfeadcffd"} Dec 02 13:21:16 crc kubenswrapper[4725]: I1202 13:21:16.389290 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7d27g" event={"ID":"154d76e0-4eac-4a29-8f3a-5aadb8762c6b","Type":"ContainerStarted","Data":"7c794093c010bbce1a0ff33d881aa4912c76b9b9d4ce39567eed657ad817c475"} Dec 02 13:21:16 crc kubenswrapper[4725]: I1202 13:21:16.411947 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-z72np" podStartSLOduration=2.371206883 podStartE2EDuration="7.411928977s" podCreationTimestamp="2025-12-02 13:21:09 +0000 UTC" firstStartedPulling="2025-12-02 13:21:10.080480587 +0000 UTC m=+1001.037122282" lastFinishedPulling="2025-12-02 13:21:15.121202671 +0000 UTC m=+1006.077844376" observedRunningTime="2025-12-02 13:21:16.406154314 +0000 UTC m=+1007.362796009" watchObservedRunningTime="2025-12-02 13:21:16.411928977 +0000 UTC m=+1007.368570672" Dec 02 13:21:17 crc kubenswrapper[4725]: I1202 13:21:17.278274 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08951dcf-7810-40b8-be36-9003fe20022b" path="/var/lib/kubelet/pods/08951dcf-7810-40b8-be36-9003fe20022b/volumes" Dec 02 13:21:17 crc kubenswrapper[4725]: I1202 13:21:17.398308 4725 generic.go:334] "Generic (PLEG): container finished" podID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerID="6e3ec9fe8664b961c59b197cc682fd6ee84f663d3385bbcb3233ec98fc495c03" exitCode=0 Dec 02 13:21:17 crc kubenswrapper[4725]: I1202 13:21:17.399017 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ede4bdcc-2588-4e91-80f7-1bcf0467e861","Type":"ContainerDied","Data":"6e3ec9fe8664b961c59b197cc682fd6ee84f663d3385bbcb3233ec98fc495c03"} Dec 02 13:21:18 crc kubenswrapper[4725]: I1202 13:21:18.430335 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ede4bdcc-2588-4e91-80f7-1bcf0467e861","Type":"ContainerStarted","Data":"375c371f1dd02d3aa09718b311330c557f3512342b53e32ba612467bf474fbdc"} Dec 02 13:21:18 crc kubenswrapper[4725]: I1202 13:21:18.431881 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:21:18 crc kubenswrapper[4725]: I1202 13:21:18.503813 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.443025132 podStartE2EDuration="1m1.503788865s" podCreationTimestamp="2025-12-02 13:20:17 +0000 UTC" firstStartedPulling="2025-12-02 13:20:19.78017461 +0000 UTC m=+950.736816305" lastFinishedPulling="2025-12-02 13:20:42.840938333 +0000 UTC m=+973.797580038" observedRunningTime="2025-12-02 13:21:18.49555959 +0000 UTC m=+1009.452201295" watchObservedRunningTime="2025-12-02 13:21:18.503788865 +0000 UTC m=+1009.460430550" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.491783 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"51144a39-c6dd-4a5f-adfb-6d62cb459485","Type":"ContainerStarted","Data":"180b34affbd1aa755e2d3aa4a0a5ed1c0e3f7b402a99e572fc04d6a5e60a1158"} Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.519559 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.876691202 podStartE2EDuration="49.519541694s" podCreationTimestamp="2025-12-02 13:20:30 +0000 UTC" firstStartedPulling="2025-12-02 13:20:44.027843587 +0000 UTC m=+974.984485282" lastFinishedPulling="2025-12-02 13:21:18.670694079 +0000 UTC m=+1009.627335774" observedRunningTime="2025-12-02 13:21:19.51695703 +0000 UTC m=+1010.473598725" watchObservedRunningTime="2025-12-02 13:21:19.519541694 +0000 UTC m=+1010.476183389" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.760476 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-xq775"] Dec 02 13:21:19 crc kubenswrapper[4725]: E1202 13:21:19.760814 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08951dcf-7810-40b8-be36-9003fe20022b" containerName="dnsmasq-dns" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.760830 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="08951dcf-7810-40b8-be36-9003fe20022b" containerName="dnsmasq-dns" Dec 02 13:21:19 crc kubenswrapper[4725]: E1202 13:21:19.760859 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08951dcf-7810-40b8-be36-9003fe20022b" containerName="init" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.760866 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="08951dcf-7810-40b8-be36-9003fe20022b" containerName="init" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.761049 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="08951dcf-7810-40b8-be36-9003fe20022b" containerName="dnsmasq-dns" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.761875 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.764208 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.782448 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-xq775"] Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.847762 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kwn9x"] Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.849027 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.851679 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.860831 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kwn9x"] Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.900802 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.900925 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.900957 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h57jv\" (UniqueName: \"kubernetes.io/projected/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-kube-api-access-h57jv\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:19 crc kubenswrapper[4725]: I1202 13:21:19.901032 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-config\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003047 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h57jv\" (UniqueName: \"kubernetes.io/projected/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-kube-api-access-h57jv\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003125 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-ovs-rundir\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003177 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg874\" (UniqueName: \"kubernetes.io/projected/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-kube-api-access-pg874\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003202 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-combined-ca-bundle\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003224 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-ovn-rundir\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003258 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-config\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003316 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003380 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003414 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-config\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.003629 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.004343 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.004602 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.004648 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-config\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.041429 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h57jv\" (UniqueName: \"kubernetes.io/projected/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-kube-api-access-h57jv\") pod \"dnsmasq-dns-6c89d5d749-xq775\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.082345 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.099230 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-xq775"] Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.104699 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-ovs-rundir\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.104762 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg874\" (UniqueName: \"kubernetes.io/projected/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-kube-api-access-pg874\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.104783 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-combined-ca-bundle\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.104800 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-ovn-rundir\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.104873 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.104894 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-config\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.106131 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-config\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.106355 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-ovs-rundir\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.109649 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-ovn-rundir\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.111683 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-combined-ca-bundle\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.121010 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.125988 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg874\" (UniqueName: \"kubernetes.io/projected/4739bfa6-9e63-4f3a-bba3-b6c21afe2a74-kube-api-access-pg874\") pod \"ovn-controller-metrics-kwn9x\" (UID: \"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74\") " pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.138836 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmdk4"] Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.141093 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.146848 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.167529 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmdk4"] Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.174748 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kwn9x" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.313074 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-dns-svc\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.313164 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-config\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.313223 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.313281 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.313328 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/33e0067a-bebb-4abf-af23-42be9e914232-kube-api-access-qx9db\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.414508 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/33e0067a-bebb-4abf-af23-42be9e914232-kube-api-access-qx9db\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.414651 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-dns-svc\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.414689 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-config\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.414740 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.414814 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.415709 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.415943 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-dns-svc\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.415975 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.416591 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-config\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.565528 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/33e0067a-bebb-4abf-af23-42be9e914232-kube-api-access-qx9db\") pod \"dnsmasq-dns-698758b865-qmdk4\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:20 crc kubenswrapper[4725]: I1202 13:21:20.567957 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:21 crc kubenswrapper[4725]: I1202 13:21:21.303829 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:21 crc kubenswrapper[4725]: E1202 13:21:21.304343 4725 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 13:21:21 crc kubenswrapper[4725]: E1202 13:21:21.304357 4725 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 13:21:21 crc kubenswrapper[4725]: E1202 13:21:21.304405 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift podName:c3fe5753-c7c5-450a-b4e6-0065deae5cd9 nodeName:}" failed. No retries permitted until 2025-12-02 13:21:37.304390309 +0000 UTC m=+1028.261032004 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift") pod "swift-storage-0" (UID: "c3fe5753-c7c5-450a-b4e6-0065deae5cd9") : configmap "swift-ring-files" not found Dec 02 13:21:21 crc kubenswrapper[4725]: I1202 13:21:21.555999 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bad22ab8-0033-4a28-b7aa-65b1c285a989","Type":"ContainerStarted","Data":"cd06c178e7dc34a9a5d3d1fb2e199c31d509296d6fb63449b7b28cf28dbbd39f"} Dec 02 13:21:21 crc kubenswrapper[4725]: I1202 13:21:21.621942 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=50.016968724 podStartE2EDuration="54.621914233s" podCreationTimestamp="2025-12-02 13:20:27 +0000 UTC" firstStartedPulling="2025-12-02 13:20:44.009735521 +0000 UTC m=+974.966377216" lastFinishedPulling="2025-12-02 13:20:48.61468103 +0000 UTC m=+979.571322725" observedRunningTime="2025-12-02 13:21:21.613423642 +0000 UTC m=+1012.570065337" watchObservedRunningTime="2025-12-02 13:21:21.621914233 +0000 UTC m=+1012.578555928" Dec 02 13:21:21 crc kubenswrapper[4725]: I1202 13:21:21.733477 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kwn9x"] Dec 02 13:21:21 crc kubenswrapper[4725]: I1202 13:21:21.853608 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmdk4"] Dec 02 13:21:21 crc kubenswrapper[4725]: I1202 13:21:21.920603 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-xq775"] Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.038192 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.040406 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.044359 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.044654 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.044798 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8h4zt" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.044912 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.054534 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125641 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/984ec540-6517-43f8-90e9-751b57ff0165-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125700 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125728 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125763 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzrnc\" (UniqueName: \"kubernetes.io/projected/984ec540-6517-43f8-90e9-751b57ff0165-kube-api-access-hzrnc\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125794 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984ec540-6517-43f8-90e9-751b57ff0165-config\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125834 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/984ec540-6517-43f8-90e9-751b57ff0165-scripts\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.125875 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.227414 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzrnc\" (UniqueName: \"kubernetes.io/projected/984ec540-6517-43f8-90e9-751b57ff0165-kube-api-access-hzrnc\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.227472 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984ec540-6517-43f8-90e9-751b57ff0165-config\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.227495 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/984ec540-6517-43f8-90e9-751b57ff0165-scripts\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.227531 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.227625 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/984ec540-6517-43f8-90e9-751b57ff0165-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.227652 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.228087 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.228852 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/984ec540-6517-43f8-90e9-751b57ff0165-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.229144 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984ec540-6517-43f8-90e9-751b57ff0165-config\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.230694 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/984ec540-6517-43f8-90e9-751b57ff0165-scripts\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.233017 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.234177 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.248810 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzrnc\" (UniqueName: \"kubernetes.io/projected/984ec540-6517-43f8-90e9-751b57ff0165-kube-api-access-hzrnc\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.250575 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/984ec540-6517-43f8-90e9-751b57ff0165-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"984ec540-6517-43f8-90e9-751b57ff0165\") " pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.353208 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.566592 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kwn9x" event={"ID":"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74","Type":"ContainerStarted","Data":"e2dabe4ff6563a56af09833c1859616b054d5b0f29133c53ccf3d4c8a9bbdf25"} Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.568122 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" event={"ID":"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c","Type":"ContainerStarted","Data":"64213cb9989940ee981d374c59f0eb3674a48c54259d74daf45e13020c21b81f"} Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.569135 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmdk4" event={"ID":"33e0067a-bebb-4abf-af23-42be9e914232","Type":"ContainerStarted","Data":"2a779da4a98fdbb55f59071416abbfc18764637bf097e2efb7d981a537664d4f"} Dec 02 13:21:22 crc kubenswrapper[4725]: I1202 13:21:22.805339 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.086866 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4ngrg" podUID="79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7" containerName="ovn-controller" probeResult="failure" output=< Dec 02 13:21:23 crc kubenswrapper[4725]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 02 13:21:23 crc kubenswrapper[4725]: > Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.162589 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.577792 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"984ec540-6517-43f8-90e9-751b57ff0165","Type":"ContainerStarted","Data":"8e4200674bcc5061c305ddb8d55ef55cbb7934d256a6e94036226c1d833220a9"} Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.580553 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kwn9x" event={"ID":"4739bfa6-9e63-4f3a-bba3-b6c21afe2a74","Type":"ContainerStarted","Data":"91ea9dcda73a0226f3c4cceaa4ee09bef21ae83359541acdc6b352aa0a460744"} Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.584734 4725 generic.go:334] "Generic (PLEG): container finished" podID="da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" containerID="bef71d1000f65f5eaf7459b0165ce66a5f99ca2c530016331bf87b423b3ec010" exitCode=0 Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.584871 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" event={"ID":"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c","Type":"ContainerDied","Data":"bef71d1000f65f5eaf7459b0165ce66a5f99ca2c530016331bf87b423b3ec010"} Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.587660 4725 generic.go:334] "Generic (PLEG): container finished" podID="33e0067a-bebb-4abf-af23-42be9e914232" containerID="d5aabec06ed4d824f780c8ec733dceecc410de95e0198229412d107b7bb213c4" exitCode=0 Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.587683 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmdk4" event={"ID":"33e0067a-bebb-4abf-af23-42be9e914232","Type":"ContainerDied","Data":"d5aabec06ed4d824f780c8ec733dceecc410de95e0198229412d107b7bb213c4"} Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.606141 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kwn9x" podStartSLOduration=4.606122337 podStartE2EDuration="4.606122337s" podCreationTimestamp="2025-12-02 13:21:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:23.597693108 +0000 UTC m=+1014.554334803" watchObservedRunningTime="2025-12-02 13:21:23.606122337 +0000 UTC m=+1014.562764032" Dec 02 13:21:23 crc kubenswrapper[4725]: I1202 13:21:23.991565 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.078174 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-dns-svc\") pod \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.078283 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h57jv\" (UniqueName: \"kubernetes.io/projected/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-kube-api-access-h57jv\") pod \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.078350 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-config\") pod \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.078557 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-ovsdbserver-sb\") pod \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\" (UID: \"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c\") " Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.085117 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-kube-api-access-h57jv" (OuterVolumeSpecName: "kube-api-access-h57jv") pod "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" (UID: "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c"). InnerVolumeSpecName "kube-api-access-h57jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.100757 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-config" (OuterVolumeSpecName: "config") pod "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" (UID: "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.108529 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" (UID: "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.116399 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" (UID: "da05e92f-d5ce-4e3f-aaaf-3115f2d0775c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.181291 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.181332 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h57jv\" (UniqueName: \"kubernetes.io/projected/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-kube-api-access-h57jv\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.181347 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.181360 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.599045 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" event={"ID":"da05e92f-d5ce-4e3f-aaaf-3115f2d0775c","Type":"ContainerDied","Data":"64213cb9989940ee981d374c59f0eb3674a48c54259d74daf45e13020c21b81f"} Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.599076 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-xq775" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.599445 4725 scope.go:117] "RemoveContainer" containerID="bef71d1000f65f5eaf7459b0165ce66a5f99ca2c530016331bf87b423b3ec010" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.601247 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmdk4" event={"ID":"33e0067a-bebb-4abf-af23-42be9e914232","Type":"ContainerStarted","Data":"a47074e6329a6a09eb6dcfff9b5be160af9ab43451062e9db24d19f7dd5aa947"} Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.601285 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.626982 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qmdk4" podStartSLOduration=4.626964983 podStartE2EDuration="4.626964983s" podCreationTimestamp="2025-12-02 13:21:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:24.623817385 +0000 UTC m=+1015.580459110" watchObservedRunningTime="2025-12-02 13:21:24.626964983 +0000 UTC m=+1015.583606678" Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.681767 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-xq775"] Dec 02 13:21:24 crc kubenswrapper[4725]: I1202 13:21:24.697677 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-xq775"] Dec 02 13:21:25 crc kubenswrapper[4725]: I1202 13:21:25.276583 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" path="/var/lib/kubelet/pods/da05e92f-d5ce-4e3f-aaaf-3115f2d0775c/volumes" Dec 02 13:21:25 crc kubenswrapper[4725]: I1202 13:21:25.614511 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"984ec540-6517-43f8-90e9-751b57ff0165","Type":"ContainerStarted","Data":"24bf586595fe5980b452933c471d601206f49473d68890507db67af8c1d68804"} Dec 02 13:21:25 crc kubenswrapper[4725]: I1202 13:21:25.614570 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"984ec540-6517-43f8-90e9-751b57ff0165","Type":"ContainerStarted","Data":"f1ee34d0c6e0b9ef4bace31fd510caecbfd01a32b82c86b095dbea661664d536"} Dec 02 13:21:25 crc kubenswrapper[4725]: I1202 13:21:25.614654 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 02 13:21:25 crc kubenswrapper[4725]: I1202 13:21:25.633252 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.013301305 podStartE2EDuration="4.633233566s" podCreationTimestamp="2025-12-02 13:21:21 +0000 UTC" firstStartedPulling="2025-12-02 13:21:22.81765377 +0000 UTC m=+1013.774295465" lastFinishedPulling="2025-12-02 13:21:24.437586031 +0000 UTC m=+1015.394227726" observedRunningTime="2025-12-02 13:21:25.630901338 +0000 UTC m=+1016.587543033" watchObservedRunningTime="2025-12-02 13:21:25.633233566 +0000 UTC m=+1016.589875261" Dec 02 13:21:26 crc kubenswrapper[4725]: I1202 13:21:26.627682 4725 generic.go:334] "Generic (PLEG): container finished" podID="dda4b826-1871-43b5-808d-45d84f2f2483" containerID="41ee4eaa3ba651f62974272941794fd14a4c40b5255439aee5823eebfeadcffd" exitCode=0 Dec 02 13:21:26 crc kubenswrapper[4725]: I1202 13:21:26.627757 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z72np" event={"ID":"dda4b826-1871-43b5-808d-45d84f2f2483","Type":"ContainerDied","Data":"41ee4eaa3ba651f62974272941794fd14a4c40b5255439aee5823eebfeadcffd"} Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.092603 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4ngrg" podUID="79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7" containerName="ovn-controller" probeResult="failure" output=< Dec 02 13:21:28 crc kubenswrapper[4725]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 02 13:21:28 crc kubenswrapper[4725]: > Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.218789 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bvlkl" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.433530 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4ngrg-config-f6zgf"] Dec 02 13:21:28 crc kubenswrapper[4725]: E1202 13:21:28.434433 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" containerName="init" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.434564 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" containerName="init" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.434861 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="da05e92f-d5ce-4e3f-aaaf-3115f2d0775c" containerName="init" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.435770 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.439709 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.444352 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ngrg-config-f6zgf"] Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.499782 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-additional-scripts\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.499856 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run-ovn\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.499881 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-log-ovn\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.499925 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-scripts\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.499986 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.500014 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bktmz\" (UniqueName: \"kubernetes.io/projected/7b7b00bc-e99b-435f-b72d-93874a114cee-kube-api-access-bktmz\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601178 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601253 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bktmz\" (UniqueName: \"kubernetes.io/projected/7b7b00bc-e99b-435f-b72d-93874a114cee-kube-api-access-bktmz\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601588 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-additional-scripts\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601635 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601648 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run-ovn\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601718 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run-ovn\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601752 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-log-ovn\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601877 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-log-ovn\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.601885 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-scripts\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.602652 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-additional-scripts\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.604169 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-scripts\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.625572 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bktmz\" (UniqueName: \"kubernetes.io/projected/7b7b00bc-e99b-435f-b72d-93874a114cee-kube-api-access-bktmz\") pod \"ovn-controller-4ngrg-config-f6zgf\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.645343 4725 generic.go:334] "Generic (PLEG): container finished" podID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerID="988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113" exitCode=0 Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.645393 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d630b03-f4dd-48ac-a72b-b457e3651c76","Type":"ContainerDied","Data":"988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113"} Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.775893 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:28 crc kubenswrapper[4725]: I1202 13:21:28.930169 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 02 13:21:30 crc kubenswrapper[4725]: I1202 13:21:30.570784 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:21:30 crc kubenswrapper[4725]: I1202 13:21:30.684205 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxp5n"] Dec 02 13:21:30 crc kubenswrapper[4725]: I1202 13:21:30.684543 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="dnsmasq-dns" containerID="cri-o://164e78819f34d4adc0d2fb9200dadd42794af848fca8c10e95ea1c7b0f5abf18" gracePeriod=10 Dec 02 13:21:31 crc kubenswrapper[4725]: I1202 13:21:31.673914 4725 generic.go:334] "Generic (PLEG): container finished" podID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerID="164e78819f34d4adc0d2fb9200dadd42794af848fca8c10e95ea1c7b0f5abf18" exitCode=0 Dec 02 13:21:31 crc kubenswrapper[4725]: I1202 13:21:31.673955 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" event={"ID":"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b","Type":"ContainerDied","Data":"164e78819f34d4adc0d2fb9200dadd42794af848fca8c10e95ea1c7b0f5abf18"} Dec 02 13:21:33 crc kubenswrapper[4725]: I1202 13:21:33.104266 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4ngrg" podUID="79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7" containerName="ovn-controller" probeResult="failure" output=< Dec 02 13:21:33 crc kubenswrapper[4725]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 02 13:21:33 crc kubenswrapper[4725]: > Dec 02 13:21:34 crc kubenswrapper[4725]: I1202 13:21:34.535163 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.113:5353: connect: connection refused" Dec 02 13:21:35 crc kubenswrapper[4725]: I1202 13:21:35.415762 4725 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod90ffa1aa_9afc_47dd_a3c0_cb1358a68ad5.slice" Dec 02 13:21:35 crc kubenswrapper[4725]: E1202 13:21:35.416033 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod90ffa1aa_9afc_47dd_a3c0_cb1358a68ad5.slice" pod="openstack/keystone-81d8-account-create-update-tvqbp" podUID="90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" Dec 02 13:21:35 crc kubenswrapper[4725]: I1202 13:21:35.721234 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-81d8-account-create-update-tvqbp" Dec 02 13:21:36 crc kubenswrapper[4725]: E1202 13:21:36.474867 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 02 13:21:36 crc kubenswrapper[4725]: E1202 13:21:36.476007 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6lsp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-7d27g_openstack(154d76e0-4eac-4a29-8f3a-5aadb8762c6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:21:36 crc kubenswrapper[4725]: E1202 13:21:36.477169 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-7d27g" podUID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.596887 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676071 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-dispersionconf\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676193 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbbdh\" (UniqueName: \"kubernetes.io/projected/dda4b826-1871-43b5-808d-45d84f2f2483-kube-api-access-cbbdh\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676290 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dda4b826-1871-43b5-808d-45d84f2f2483-etc-swift\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676434 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-scripts\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676614 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-ring-data-devices\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676688 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-swiftconf\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.676711 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-combined-ca-bundle\") pod \"dda4b826-1871-43b5-808d-45d84f2f2483\" (UID: \"dda4b826-1871-43b5-808d-45d84f2f2483\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.677886 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.678104 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dda4b826-1871-43b5-808d-45d84f2f2483-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.708986 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda4b826-1871-43b5-808d-45d84f2f2483-kube-api-access-cbbdh" (OuterVolumeSpecName: "kube-api-access-cbbdh") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "kube-api-access-cbbdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.709006 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.713681 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.734874 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z72np" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.735212 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z72np" event={"ID":"dda4b826-1871-43b5-808d-45d84f2f2483","Type":"ContainerDied","Data":"55b6b19ddd9dc165b7cef63170558a4998fce79bd4f646f2ecf30bebbe5d08c3"} Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.735256 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55b6b19ddd9dc165b7cef63170558a4998fce79bd4f646f2ecf30bebbe5d08c3" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.739630 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-scripts" (OuterVolumeSpecName: "scripts") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.743536 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d630b03-f4dd-48ac-a72b-b457e3651c76","Type":"ContainerStarted","Data":"4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f"} Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.743917 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 02 13:21:36 crc kubenswrapper[4725]: E1202 13:21:36.745303 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-7d27g" podUID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.763696 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dda4b826-1871-43b5-808d-45d84f2f2483" (UID: "dda4b826-1871-43b5-808d-45d84f2f2483"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779641 4725 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dda4b826-1871-43b5-808d-45d84f2f2483-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779680 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779691 4725 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dda4b826-1871-43b5-808d-45d84f2f2483-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779701 4725 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779710 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779719 4725 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dda4b826-1871-43b5-808d-45d84f2f2483-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.779736 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbbdh\" (UniqueName: \"kubernetes.io/projected/dda4b826-1871-43b5-808d-45d84f2f2483-kube-api-access-cbbdh\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.798255 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371957.056543 podStartE2EDuration="1m19.798232864s" podCreationTimestamp="2025-12-02 13:20:17 +0000 UTC" firstStartedPulling="2025-12-02 13:20:19.041589297 +0000 UTC m=+949.998230992" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:36.788099072 +0000 UTC m=+1027.744740767" watchObservedRunningTime="2025-12-02 13:21:36.798232864 +0000 UTC m=+1027.754874569" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.817737 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.880778 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-config\") pod \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.880832 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc\") pod \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.880972 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cdrl\" (UniqueName: \"kubernetes.io/projected/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-kube-api-access-7cdrl\") pod \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.883871 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-kube-api-access-7cdrl" (OuterVolumeSpecName: "kube-api-access-7cdrl") pod "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" (UID: "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b"). InnerVolumeSpecName "kube-api-access-7cdrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: E1202 13:21:36.914526 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc podName:0a0325e3-d5eb-4a03-aa3a-bbc172697b9b nodeName:}" failed. No retries permitted until 2025-12-02 13:21:37.414365897 +0000 UTC m=+1028.371007592 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc") pod "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" (UID: "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b") : error deleting /var/lib/kubelet/pods/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b/volume-subpaths: remove /var/lib/kubelet/pods/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b/volume-subpaths: no such file or directory Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.914799 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-config" (OuterVolumeSpecName: "config") pod "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" (UID: "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.983045 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cdrl\" (UniqueName: \"kubernetes.io/projected/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-kube-api-access-7cdrl\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.983268 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:36 crc kubenswrapper[4725]: I1202 13:21:36.989347 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4ngrg-config-f6zgf"] Dec 02 13:21:36 crc kubenswrapper[4725]: W1202 13:21:36.992714 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b7b00bc_e99b_435f_b72d_93874a114cee.slice/crio-3f016971811dd27a4249834959425fe8db9a9a35109d05820275add624868f13 WatchSource:0}: Error finding container 3f016971811dd27a4249834959425fe8db9a9a35109d05820275add624868f13: Status 404 returned error can't find the container with id 3f016971811dd27a4249834959425fe8db9a9a35109d05820275add624868f13 Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.389614 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.398272 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c3fe5753-c7c5-450a-b4e6-0065deae5cd9-etc-swift\") pod \"swift-storage-0\" (UID: \"c3fe5753-c7c5-450a-b4e6-0065deae5cd9\") " pod="openstack/swift-storage-0" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.429515 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.439754 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.491675 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc\") pod \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\" (UID: \"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b\") " Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.492632 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" (UID: "0a0325e3-d5eb-4a03-aa3a-bbc172697b9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.594640 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.758579 4725 generic.go:334] "Generic (PLEG): container finished" podID="7b7b00bc-e99b-435f-b72d-93874a114cee" containerID="115389f2abd70000c9b731354b036fa08a2b1652d50bb65ca5e5328213ac48ea" exitCode=0 Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.758706 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ngrg-config-f6zgf" event={"ID":"7b7b00bc-e99b-435f-b72d-93874a114cee","Type":"ContainerDied","Data":"115389f2abd70000c9b731354b036fa08a2b1652d50bb65ca5e5328213ac48ea"} Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.758740 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ngrg-config-f6zgf" event={"ID":"7b7b00bc-e99b-435f-b72d-93874a114cee","Type":"ContainerStarted","Data":"3f016971811dd27a4249834959425fe8db9a9a35109d05820275add624868f13"} Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.763100 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" event={"ID":"0a0325e3-d5eb-4a03-aa3a-bbc172697b9b","Type":"ContainerDied","Data":"6d5c5d52eb2247b600ffb5fa113b7bc2ddb25d5ffb0b01f616c7620f8b4fdad1"} Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.763149 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-fxp5n" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.763168 4725 scope.go:117] "RemoveContainer" containerID="164e78819f34d4adc0d2fb9200dadd42794af848fca8c10e95ea1c7b0f5abf18" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.797207 4725 scope.go:117] "RemoveContainer" containerID="226de640a1f4a608f4933ef43c8ea968e0a3d8e31ac183e3b803a7cd41ea4967" Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.813007 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxp5n"] Dec 02 13:21:37 crc kubenswrapper[4725]: I1202 13:21:37.822398 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-fxp5n"] Dec 02 13:21:38 crc kubenswrapper[4725]: I1202 13:21:38.033524 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 02 13:21:38 crc kubenswrapper[4725]: W1202 13:21:38.040628 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3fe5753_c7c5_450a_b4e6_0065deae5cd9.slice/crio-ade35927b9965162378ea11370a5f0b94a611683ae4da94b2468294b04d09ef4 WatchSource:0}: Error finding container ade35927b9965162378ea11370a5f0b94a611683ae4da94b2468294b04d09ef4: Status 404 returned error can't find the container with id ade35927b9965162378ea11370a5f0b94a611683ae4da94b2468294b04d09ef4 Dec 02 13:21:38 crc kubenswrapper[4725]: I1202 13:21:38.077890 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4ngrg" Dec 02 13:21:38 crc kubenswrapper[4725]: I1202 13:21:38.768713 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"ade35927b9965162378ea11370a5f0b94a611683ae4da94b2468294b04d09ef4"} Dec 02 13:21:38 crc kubenswrapper[4725]: I1202 13:21:38.937601 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.129977 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.219952 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bktmz\" (UniqueName: \"kubernetes.io/projected/7b7b00bc-e99b-435f-b72d-93874a114cee-kube-api-access-bktmz\") pod \"7b7b00bc-e99b-435f-b72d-93874a114cee\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220085 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-scripts\") pod \"7b7b00bc-e99b-435f-b72d-93874a114cee\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220145 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-additional-scripts\") pod \"7b7b00bc-e99b-435f-b72d-93874a114cee\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220219 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-log-ovn\") pod \"7b7b00bc-e99b-435f-b72d-93874a114cee\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220262 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run-ovn\") pod \"7b7b00bc-e99b-435f-b72d-93874a114cee\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220334 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run\") pod \"7b7b00bc-e99b-435f-b72d-93874a114cee\" (UID: \"7b7b00bc-e99b-435f-b72d-93874a114cee\") " Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220335 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7b7b00bc-e99b-435f-b72d-93874a114cee" (UID: "7b7b00bc-e99b-435f-b72d-93874a114cee"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220397 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7b7b00bc-e99b-435f-b72d-93874a114cee" (UID: "7b7b00bc-e99b-435f-b72d-93874a114cee"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220489 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run" (OuterVolumeSpecName: "var-run") pod "7b7b00bc-e99b-435f-b72d-93874a114cee" (UID: "7b7b00bc-e99b-435f-b72d-93874a114cee"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220804 4725 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220831 4725 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220843 4725 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7b7b00bc-e99b-435f-b72d-93874a114cee-var-run\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.220846 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7b7b00bc-e99b-435f-b72d-93874a114cee" (UID: "7b7b00bc-e99b-435f-b72d-93874a114cee"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.221179 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-scripts" (OuterVolumeSpecName: "scripts") pod "7b7b00bc-e99b-435f-b72d-93874a114cee" (UID: "7b7b00bc-e99b-435f-b72d-93874a114cee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.244619 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b7b00bc-e99b-435f-b72d-93874a114cee-kube-api-access-bktmz" (OuterVolumeSpecName: "kube-api-access-bktmz") pod "7b7b00bc-e99b-435f-b72d-93874a114cee" (UID: "7b7b00bc-e99b-435f-b72d-93874a114cee"). InnerVolumeSpecName "kube-api-access-bktmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.282800 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" path="/var/lib/kubelet/pods/0a0325e3-d5eb-4a03-aa3a-bbc172697b9b/volumes" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.325014 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.325054 4725 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7b7b00bc-e99b-435f-b72d-93874a114cee-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.325067 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bktmz\" (UniqueName: \"kubernetes.io/projected/7b7b00bc-e99b-435f-b72d-93874a114cee-kube-api-access-bktmz\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.783567 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4ngrg-config-f6zgf" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.783573 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4ngrg-config-f6zgf" event={"ID":"7b7b00bc-e99b-435f-b72d-93874a114cee","Type":"ContainerDied","Data":"3f016971811dd27a4249834959425fe8db9a9a35109d05820275add624868f13"} Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.784346 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f016971811dd27a4249834959425fe8db9a9a35109d05820275add624868f13" Dec 02 13:21:39 crc kubenswrapper[4725]: I1202 13:21:39.784967 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"818a60fbd3c6fc0748e6506abb5256c8676163ff83bb71181b7e276feea665a4"} Dec 02 13:21:40 crc kubenswrapper[4725]: I1202 13:21:40.304248 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4ngrg-config-f6zgf"] Dec 02 13:21:40 crc kubenswrapper[4725]: I1202 13:21:40.320172 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4ngrg-config-f6zgf"] Dec 02 13:21:40 crc kubenswrapper[4725]: I1202 13:21:40.798305 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"20458667006cbdb76459e826601e09130f5ae4d428a47a641861d6cf81a8b52f"} Dec 02 13:21:40 crc kubenswrapper[4725]: I1202 13:21:40.799946 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"cbd696d771deaa9b89c5b2fa89a4d246063f91a61d0b10cf606906f5fa1f3070"} Dec 02 13:21:40 crc kubenswrapper[4725]: I1202 13:21:40.800206 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"c1137a6b37325be56adf341ac9114be1e1638252c8fa41d6d92eb833f20d72c6"} Dec 02 13:21:41 crc kubenswrapper[4725]: I1202 13:21:41.279028 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b7b00bc-e99b-435f-b72d-93874a114cee" path="/var/lib/kubelet/pods/7b7b00bc-e99b-435f-b72d-93874a114cee/volumes" Dec 02 13:21:42 crc kubenswrapper[4725]: I1202 13:21:42.880235 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"0437eb698e4300275e8251c9d71bb34b5d675ba7b83ebd8249a1ad6f515b929f"} Dec 02 13:21:42 crc kubenswrapper[4725]: I1202 13:21:42.880761 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"e30e2a22c21c9d0319a9673d905452709a41b557b786e78721dceb4d8e3140e9"} Dec 02 13:21:42 crc kubenswrapper[4725]: I1202 13:21:42.880772 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"c4f86cdadf2ee243b69d39163a1aa37c849e10f339ecf81c3e88c7fac0584597"} Dec 02 13:21:43 crc kubenswrapper[4725]: I1202 13:21:43.891939 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"9004441319b86658a056f0e445d7e2671d4f3939b7a8c76f9994e4cbe6c63a80"} Dec 02 13:21:44 crc kubenswrapper[4725]: I1202 13:21:44.902831 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"a8e5c8242e7f22ae62df1e9a1407f9e20335771a15f3ee1d875aea66267cfcde"} Dec 02 13:21:46 crc kubenswrapper[4725]: I1202 13:21:45.918765 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"dc395fce17a7e9cb8d48322a19621d34c5e0950e504c09b9bc23affa676cc8c2"} Dec 02 13:21:46 crc kubenswrapper[4725]: I1202 13:21:45.919108 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"48154e53275916533f28ff8ffb600c7701b459511a9710b7b5c4087320f817df"} Dec 02 13:21:46 crc kubenswrapper[4725]: I1202 13:21:46.932484 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"5dbc6d5783d9a168a53c7cd8c17c6a673fc8a88d6dfef1507423e3b2cab3aa95"} Dec 02 13:21:47 crc kubenswrapper[4725]: I1202 13:21:47.948732 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"5287b4e11833d5325f451bf900ea24dccbe71142c8973c6d8b8f2fbcffb6e4bf"} Dec 02 13:21:47 crc kubenswrapper[4725]: I1202 13:21:47.949075 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"a3c44bc23d7081d40dfe0f5df740b612d778b7c3f8ca71dcb8c7c980ab598186"} Dec 02 13:21:48 crc kubenswrapper[4725]: I1202 13:21:48.612886 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.008651 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-5cvgg"] Dec 02 13:21:49 crc kubenswrapper[4725]: E1202 13:21:49.009235 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda4b826-1871-43b5-808d-45d84f2f2483" containerName="swift-ring-rebalance" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009252 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda4b826-1871-43b5-808d-45d84f2f2483" containerName="swift-ring-rebalance" Dec 02 13:21:49 crc kubenswrapper[4725]: E1202 13:21:49.009265 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="init" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009272 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="init" Dec 02 13:21:49 crc kubenswrapper[4725]: E1202 13:21:49.009283 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b7b00bc-e99b-435f-b72d-93874a114cee" containerName="ovn-config" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009288 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b7b00bc-e99b-435f-b72d-93874a114cee" containerName="ovn-config" Dec 02 13:21:49 crc kubenswrapper[4725]: E1202 13:21:49.009304 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="dnsmasq-dns" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009310 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="dnsmasq-dns" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009496 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a0325e3-d5eb-4a03-aa3a-bbc172697b9b" containerName="dnsmasq-dns" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009519 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="dda4b826-1871-43b5-808d-45d84f2f2483" containerName="swift-ring-rebalance" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.009541 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b7b00bc-e99b-435f-b72d-93874a114cee" containerName="ovn-config" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.010020 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.044637 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3607fd3f-e379-42ee-b778-5d1540371a35-operator-scripts\") pod \"cinder-db-create-5cvgg\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.044989 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlkl5\" (UniqueName: \"kubernetes.io/projected/3607fd3f-e379-42ee-b778-5d1540371a35-kube-api-access-tlkl5\") pod \"cinder-db-create-5cvgg\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.051649 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5cvgg"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.136199 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3f8e-account-create-update-wcchj"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.137765 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.141947 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.152587 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8bl2c"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.153086 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-operator-scripts\") pod \"cinder-3f8e-account-create-update-wcchj\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.153186 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3607fd3f-e379-42ee-b778-5d1540371a35-operator-scripts\") pod \"cinder-db-create-5cvgg\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.153222 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlkl5\" (UniqueName: \"kubernetes.io/projected/3607fd3f-e379-42ee-b778-5d1540371a35-kube-api-access-tlkl5\") pod \"cinder-db-create-5cvgg\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.153247 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4tzl\" (UniqueName: \"kubernetes.io/projected/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-kube-api-access-h4tzl\") pod \"cinder-3f8e-account-create-update-wcchj\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.154151 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3607fd3f-e379-42ee-b778-5d1540371a35-operator-scripts\") pod \"cinder-db-create-5cvgg\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.154371 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.170584 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8bl2c"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.181661 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3f8e-account-create-update-wcchj"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.202908 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlkl5\" (UniqueName: \"kubernetes.io/projected/3607fd3f-e379-42ee-b778-5d1540371a35-kube-api-access-tlkl5\") pod \"cinder-db-create-5cvgg\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.257757 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slc7c\" (UniqueName: \"kubernetes.io/projected/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-kube-api-access-slc7c\") pod \"barbican-db-create-8bl2c\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.257843 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4tzl\" (UniqueName: \"kubernetes.io/projected/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-kube-api-access-h4tzl\") pod \"cinder-3f8e-account-create-update-wcchj\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.257894 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-operator-scripts\") pod \"barbican-db-create-8bl2c\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.257930 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-operator-scripts\") pod \"cinder-3f8e-account-create-update-wcchj\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.258797 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-operator-scripts\") pod \"cinder-3f8e-account-create-update-wcchj\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.286622 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4tzl\" (UniqueName: \"kubernetes.io/projected/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-kube-api-access-h4tzl\") pod \"cinder-3f8e-account-create-update-wcchj\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.359761 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slc7c\" (UniqueName: \"kubernetes.io/projected/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-kube-api-access-slc7c\") pod \"barbican-db-create-8bl2c\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.360215 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-operator-scripts\") pod \"barbican-db-create-8bl2c\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.361139 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-operator-scripts\") pod \"barbican-db-create-8bl2c\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.374899 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.393405 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slc7c\" (UniqueName: \"kubernetes.io/projected/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-kube-api-access-slc7c\") pod \"barbican-db-create-8bl2c\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.408446 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tghwc"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.410167 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.441429 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tghwc"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.449097 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e6bf-account-create-update-hhhqj"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.450788 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.454182 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.459387 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zmwkf"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.460949 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.463994 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwxj4\" (UniqueName: \"kubernetes.io/projected/b3946847-5d4a-4fcc-acac-7a49400db5b2-kube-api-access-rwxj4\") pod \"barbican-e6bf-account-create-update-hhhqj\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.464084 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3946847-5d4a-4fcc-acac-7a49400db5b2-operator-scripts\") pod \"barbican-e6bf-account-create-update-hhhqj\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.464243 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fgvq\" (UniqueName: \"kubernetes.io/projected/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-kube-api-access-9fgvq\") pod \"neutron-db-create-tghwc\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.464299 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-operator-scripts\") pod \"neutron-db-create-tghwc\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.466082 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.466490 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.466923 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.471527 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-twszr" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.471947 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.473616 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e6bf-account-create-update-hhhqj"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.496411 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.531635 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zmwkf"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.558139 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3a9a-account-create-update-wkmbs"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.559878 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.562229 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.566731 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-operator-scripts\") pod \"neutron-db-create-tghwc\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.566850 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwxj4\" (UniqueName: \"kubernetes.io/projected/b3946847-5d4a-4fcc-acac-7a49400db5b2-kube-api-access-rwxj4\") pod \"barbican-e6bf-account-create-update-hhhqj\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.566903 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-combined-ca-bundle\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.566937 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3946847-5d4a-4fcc-acac-7a49400db5b2-operator-scripts\") pod \"barbican-e6bf-account-create-update-hhhqj\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.566959 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-config-data\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.567060 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fgvq\" (UniqueName: \"kubernetes.io/projected/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-kube-api-access-9fgvq\") pod \"neutron-db-create-tghwc\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.567090 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fhjd\" (UniqueName: \"kubernetes.io/projected/c76d8c86-54fe-4bb2-9590-76e2a56d8659-kube-api-access-7fhjd\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.568443 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-operator-scripts\") pod \"neutron-db-create-tghwc\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.569563 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3946847-5d4a-4fcc-acac-7a49400db5b2-operator-scripts\") pod \"barbican-e6bf-account-create-update-hhhqj\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.576680 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3a9a-account-create-update-wkmbs"] Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.598239 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwxj4\" (UniqueName: \"kubernetes.io/projected/b3946847-5d4a-4fcc-acac-7a49400db5b2-kube-api-access-rwxj4\") pod \"barbican-e6bf-account-create-update-hhhqj\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.614539 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fgvq\" (UniqueName: \"kubernetes.io/projected/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-kube-api-access-9fgvq\") pod \"neutron-db-create-tghwc\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.669137 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-combined-ca-bundle\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.669213 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-config-data\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.669352 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fhjd\" (UniqueName: \"kubernetes.io/projected/c76d8c86-54fe-4bb2-9590-76e2a56d8659-kube-api-access-7fhjd\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.669399 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-operator-scripts\") pod \"neutron-3a9a-account-create-update-wkmbs\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.669425 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrfsz\" (UniqueName: \"kubernetes.io/projected/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-kube-api-access-nrfsz\") pod \"neutron-3a9a-account-create-update-wkmbs\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.673320 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-combined-ca-bundle\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.674411 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-config-data\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.691708 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fhjd\" (UniqueName: \"kubernetes.io/projected/c76d8c86-54fe-4bb2-9590-76e2a56d8659-kube-api-access-7fhjd\") pod \"keystone-db-sync-zmwkf\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.757581 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.770328 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-operator-scripts\") pod \"neutron-3a9a-account-create-update-wkmbs\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.770396 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrfsz\" (UniqueName: \"kubernetes.io/projected/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-kube-api-access-nrfsz\") pod \"neutron-3a9a-account-create-update-wkmbs\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.771796 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-operator-scripts\") pod \"neutron-3a9a-account-create-update-wkmbs\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.778343 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.789388 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.789511 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrfsz\" (UniqueName: \"kubernetes.io/projected/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-kube-api-access-nrfsz\") pod \"neutron-3a9a-account-create-update-wkmbs\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:49 crc kubenswrapper[4725]: I1202 13:21:49.884920 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:50 crc kubenswrapper[4725]: I1202 13:21:50.931825 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5cvgg"] Dec 02 13:21:50 crc kubenswrapper[4725]: W1202 13:21:50.986062 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3607fd3f_e379_42ee_b778_5d1540371a35.slice/crio-c75d542714faf7b86b82702669e39ba356f291f4698d8dd98f7c067d74bfd651 WatchSource:0}: Error finding container c75d542714faf7b86b82702669e39ba356f291f4698d8dd98f7c067d74bfd651: Status 404 returned error can't find the container with id c75d542714faf7b86b82702669e39ba356f291f4698d8dd98f7c067d74bfd651 Dec 02 13:21:51 crc kubenswrapper[4725]: W1202 13:21:51.307095 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10ca0cc7_f32f_40e8_9ce9_5a0d0efa1019.slice/crio-2e672a1dded776e29e0bab7b213064d730fdd1545a03fbb5ffb80900119d6dd2 WatchSource:0}: Error finding container 2e672a1dded776e29e0bab7b213064d730fdd1545a03fbb5ffb80900119d6dd2: Status 404 returned error can't find the container with id 2e672a1dded776e29e0bab7b213064d730fdd1545a03fbb5ffb80900119d6dd2 Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.338057 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3f8e-account-create-update-wcchj"] Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.450729 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tghwc"] Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.476805 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8bl2c"] Dec 02 13:21:51 crc kubenswrapper[4725]: W1202 13:21:51.477366 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cddd3b3_e14f_4dba_a7f9_7ac6ae0bafa8.slice/crio-a77006c8d435b1a1bf2d9a6f74127e5ad71393f5ebed72566b035fb4fc3677c5 WatchSource:0}: Error finding container a77006c8d435b1a1bf2d9a6f74127e5ad71393f5ebed72566b035fb4fc3677c5: Status 404 returned error can't find the container with id a77006c8d435b1a1bf2d9a6f74127e5ad71393f5ebed72566b035fb4fc3677c5 Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.576485 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3a9a-account-create-update-wkmbs"] Dec 02 13:21:51 crc kubenswrapper[4725]: W1202 13:21:51.594743 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3946847_5d4a_4fcc_acac_7a49400db5b2.slice/crio-a5acbfe665396b707daf03c8ed53e6cb10e5ea907d08f74e929fce54a19bed2c WatchSource:0}: Error finding container a5acbfe665396b707daf03c8ed53e6cb10e5ea907d08f74e929fce54a19bed2c: Status 404 returned error can't find the container with id a5acbfe665396b707daf03c8ed53e6cb10e5ea907d08f74e929fce54a19bed2c Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.598782 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e6bf-account-create-update-hhhqj"] Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.608342 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zmwkf"] Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.983116 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3f8e-account-create-update-wcchj" event={"ID":"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019","Type":"ContainerStarted","Data":"2e672a1dded776e29e0bab7b213064d730fdd1545a03fbb5ffb80900119d6dd2"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.985718 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a9a-account-create-update-wkmbs" event={"ID":"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5","Type":"ContainerStarted","Data":"e1a628004bc1dbabf6244b90b48f7b1987f7fca2663cda9faf62e2d61c9d2499"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.987963 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e6bf-account-create-update-hhhqj" event={"ID":"b3946847-5d4a-4fcc-acac-7a49400db5b2","Type":"ContainerStarted","Data":"a5acbfe665396b707daf03c8ed53e6cb10e5ea907d08f74e929fce54a19bed2c"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.992097 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8bl2c" event={"ID":"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8","Type":"ContainerStarted","Data":"a77006c8d435b1a1bf2d9a6f74127e5ad71393f5ebed72566b035fb4fc3677c5"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.995754 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tghwc" event={"ID":"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7","Type":"ContainerStarted","Data":"e4aa8760923ff895f626acfdfdd9d83423ac9ad87f6a854c00064b8698bd291c"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.997062 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5cvgg" event={"ID":"3607fd3f-e379-42ee-b778-5d1540371a35","Type":"ContainerStarted","Data":"b85444a496bd38286c3f1aecb2e9905fc40447f2bc1e19cd6485fb817cad8caa"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.997088 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5cvgg" event={"ID":"3607fd3f-e379-42ee-b778-5d1540371a35","Type":"ContainerStarted","Data":"c75d542714faf7b86b82702669e39ba356f291f4698d8dd98f7c067d74bfd651"} Dec 02 13:21:51 crc kubenswrapper[4725]: I1202 13:21:51.999147 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zmwkf" event={"ID":"c76d8c86-54fe-4bb2-9590-76e2a56d8659","Type":"ContainerStarted","Data":"f417cc723ab481c39589937b86dcc63bfed7d8d852443198a6135c591f79ac30"} Dec 02 13:21:52 crc kubenswrapper[4725]: I1202 13:21:52.019266 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-5cvgg" podStartSLOduration=4.019246146 podStartE2EDuration="4.019246146s" podCreationTimestamp="2025-12-02 13:21:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:52.011406951 +0000 UTC m=+1042.968048646" watchObservedRunningTime="2025-12-02 13:21:52.019246146 +0000 UTC m=+1042.975887841" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.010954 4725 generic.go:334] "Generic (PLEG): container finished" podID="b3946847-5d4a-4fcc-acac-7a49400db5b2" containerID="a4e89f6e9c6fa2fe3742e3ceaa282dd38e3e0396a9f1e04f6b7d623789005d36" exitCode=0 Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.011397 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e6bf-account-create-update-hhhqj" event={"ID":"b3946847-5d4a-4fcc-acac-7a49400db5b2","Type":"ContainerDied","Data":"a4e89f6e9c6fa2fe3742e3ceaa282dd38e3e0396a9f1e04f6b7d623789005d36"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.012865 4725 generic.go:334] "Generic (PLEG): container finished" podID="1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" containerID="1e9f5e0868aa10e463afac90fa39f6c8ae20b9ffa2bc874ff0e970d7f0874046" exitCode=0 Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.012929 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8bl2c" event={"ID":"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8","Type":"ContainerDied","Data":"1e9f5e0868aa10e463afac90fa39f6c8ae20b9ffa2bc874ff0e970d7f0874046"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.014825 4725 generic.go:334] "Generic (PLEG): container finished" podID="5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" containerID="28b54163a4f9afb5c4fb334e58ae4e155c4241896fe8be52b62a553a49de9d86" exitCode=0 Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.014901 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tghwc" event={"ID":"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7","Type":"ContainerDied","Data":"28b54163a4f9afb5c4fb334e58ae4e155c4241896fe8be52b62a553a49de9d86"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.016155 4725 generic.go:334] "Generic (PLEG): container finished" podID="3607fd3f-e379-42ee-b778-5d1540371a35" containerID="b85444a496bd38286c3f1aecb2e9905fc40447f2bc1e19cd6485fb817cad8caa" exitCode=0 Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.016204 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5cvgg" event={"ID":"3607fd3f-e379-42ee-b778-5d1540371a35","Type":"ContainerDied","Data":"b85444a496bd38286c3f1aecb2e9905fc40447f2bc1e19cd6485fb817cad8caa"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.024209 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c3fe5753-c7c5-450a-b4e6-0065deae5cd9","Type":"ContainerStarted","Data":"3766447cd0218daa07993ac19be066013000e5930a29b51582268859bcf90021"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.026702 4725 generic.go:334] "Generic (PLEG): container finished" podID="10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" containerID="0f9f97eecdfc4c2d1cca68aba5c072eb9b014f2a535efb2d918522dbc660ccf4" exitCode=0 Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.026746 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3f8e-account-create-update-wcchj" event={"ID":"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019","Type":"ContainerDied","Data":"0f9f97eecdfc4c2d1cca68aba5c072eb9b014f2a535efb2d918522dbc660ccf4"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.027754 4725 generic.go:334] "Generic (PLEG): container finished" podID="1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" containerID="3f90d3ab25f62602b6bc226bb3cf96e98918e5370b9e3e7a2bb566c8fb637d61" exitCode=0 Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.027781 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a9a-account-create-update-wkmbs" event={"ID":"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5","Type":"ContainerDied","Data":"3f90d3ab25f62602b6bc226bb3cf96e98918e5370b9e3e7a2bb566c8fb637d61"} Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.229102 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=42.718806965 podStartE2EDuration="49.22731321s" podCreationTimestamp="2025-12-02 13:21:04 +0000 UTC" firstStartedPulling="2025-12-02 13:21:38.043888691 +0000 UTC m=+1029.000530386" lastFinishedPulling="2025-12-02 13:21:44.552394936 +0000 UTC m=+1035.509036631" observedRunningTime="2025-12-02 13:21:53.217669431 +0000 UTC m=+1044.174311126" watchObservedRunningTime="2025-12-02 13:21:53.22731321 +0000 UTC m=+1044.183954905" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.543214 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fcphz"] Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.545618 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.549881 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.561016 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fcphz"] Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.569065 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.569110 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-config\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.569155 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz5dl\" (UniqueName: \"kubernetes.io/projected/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-kube-api-access-hz5dl\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.569171 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-svc\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.569208 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.569246 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.670784 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.670926 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.670963 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-config\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.671019 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz5dl\" (UniqueName: \"kubernetes.io/projected/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-kube-api-access-hz5dl\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.671044 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-svc\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.671092 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.672174 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.672370 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-svc\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.672538 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.672740 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.673276 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-config\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.700641 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz5dl\" (UniqueName: \"kubernetes.io/projected/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-kube-api-access-hz5dl\") pod \"dnsmasq-dns-764c5664d7-fcphz\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:53 crc kubenswrapper[4725]: I1202 13:21:53.875854 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.065925 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7d27g" event={"ID":"154d76e0-4eac-4a29-8f3a-5aadb8762c6b","Type":"ContainerStarted","Data":"6a99a7a44b1ca76d56407724b6adb597b9c0ff0392cc4225b7f17397752169c4"} Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.089058 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-7d27g" podStartSLOduration=5.7664280770000005 podStartE2EDuration="42.089039155s" podCreationTimestamp="2025-12-02 13:21:12 +0000 UTC" firstStartedPulling="2025-12-02 13:21:15.713479306 +0000 UTC m=+1006.670121001" lastFinishedPulling="2025-12-02 13:21:52.036090384 +0000 UTC m=+1042.992732079" observedRunningTime="2025-12-02 13:21:54.084652006 +0000 UTC m=+1045.041293701" watchObservedRunningTime="2025-12-02 13:21:54.089039155 +0000 UTC m=+1045.045680850" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.851762 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.861223 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.872505 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.879975 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.886646 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.897165 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-operator-scripts\") pod \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.897295 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlkl5\" (UniqueName: \"kubernetes.io/projected/3607fd3f-e379-42ee-b778-5d1540371a35-kube-api-access-tlkl5\") pod \"3607fd3f-e379-42ee-b778-5d1540371a35\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.897944 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" (UID: "5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.898522 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slc7c\" (UniqueName: \"kubernetes.io/projected/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-kube-api-access-slc7c\") pod \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.898605 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwxj4\" (UniqueName: \"kubernetes.io/projected/b3946847-5d4a-4fcc-acac-7a49400db5b2-kube-api-access-rwxj4\") pod \"b3946847-5d4a-4fcc-acac-7a49400db5b2\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.898632 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fgvq\" (UniqueName: \"kubernetes.io/projected/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-kube-api-access-9fgvq\") pod \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\" (UID: \"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.898671 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-operator-scripts\") pod \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\" (UID: \"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.898740 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3607fd3f-e379-42ee-b778-5d1540371a35-operator-scripts\") pod \"3607fd3f-e379-42ee-b778-5d1540371a35\" (UID: \"3607fd3f-e379-42ee-b778-5d1540371a35\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.898789 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3946847-5d4a-4fcc-acac-7a49400db5b2-operator-scripts\") pod \"b3946847-5d4a-4fcc-acac-7a49400db5b2\" (UID: \"b3946847-5d4a-4fcc-acac-7a49400db5b2\") " Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.899365 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.900033 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3946847-5d4a-4fcc-acac-7a49400db5b2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3946847-5d4a-4fcc-acac-7a49400db5b2" (UID: "b3946847-5d4a-4fcc-acac-7a49400db5b2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.900941 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" (UID: "1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.904345 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3607fd3f-e379-42ee-b778-5d1540371a35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3607fd3f-e379-42ee-b778-5d1540371a35" (UID: "3607fd3f-e379-42ee-b778-5d1540371a35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.913004 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3607fd3f-e379-42ee-b778-5d1540371a35-kube-api-access-tlkl5" (OuterVolumeSpecName: "kube-api-access-tlkl5") pod "3607fd3f-e379-42ee-b778-5d1540371a35" (UID: "3607fd3f-e379-42ee-b778-5d1540371a35"). InnerVolumeSpecName "kube-api-access-tlkl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.913119 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-kube-api-access-9fgvq" (OuterVolumeSpecName: "kube-api-access-9fgvq") pod "5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" (UID: "5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7"). InnerVolumeSpecName "kube-api-access-9fgvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.913214 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-kube-api-access-slc7c" (OuterVolumeSpecName: "kube-api-access-slc7c") pod "1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" (UID: "1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8"). InnerVolumeSpecName "kube-api-access-slc7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.916769 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.917611 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3946847-5d4a-4fcc-acac-7a49400db5b2-kube-api-access-rwxj4" (OuterVolumeSpecName: "kube-api-access-rwxj4") pod "b3946847-5d4a-4fcc-acac-7a49400db5b2" (UID: "b3946847-5d4a-4fcc-acac-7a49400db5b2"). InnerVolumeSpecName "kube-api-access-rwxj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:54 crc kubenswrapper[4725]: I1202 13:21:54.943156 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fcphz"] Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.001125 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4tzl\" (UniqueName: \"kubernetes.io/projected/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-kube-api-access-h4tzl\") pod \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.001281 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrfsz\" (UniqueName: \"kubernetes.io/projected/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-kube-api-access-nrfsz\") pod \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.001318 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-operator-scripts\") pod \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\" (UID: \"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5\") " Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.001402 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-operator-scripts\") pod \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\" (UID: \"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019\") " Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002204 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlkl5\" (UniqueName: \"kubernetes.io/projected/3607fd3f-e379-42ee-b778-5d1540371a35-kube-api-access-tlkl5\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002228 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slc7c\" (UniqueName: \"kubernetes.io/projected/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-kube-api-access-slc7c\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002239 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwxj4\" (UniqueName: \"kubernetes.io/projected/b3946847-5d4a-4fcc-acac-7a49400db5b2-kube-api-access-rwxj4\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002250 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fgvq\" (UniqueName: \"kubernetes.io/projected/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7-kube-api-access-9fgvq\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002264 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002275 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3607fd3f-e379-42ee-b778-5d1540371a35-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002286 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3946847-5d4a-4fcc-acac-7a49400db5b2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002269 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" (UID: "10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.002267 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" (UID: "1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.013645 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-kube-api-access-nrfsz" (OuterVolumeSpecName: "kube-api-access-nrfsz") pod "1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" (UID: "1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5"). InnerVolumeSpecName "kube-api-access-nrfsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.018675 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-kube-api-access-h4tzl" (OuterVolumeSpecName: "kube-api-access-h4tzl") pod "10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" (UID: "10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019"). InnerVolumeSpecName "kube-api-access-h4tzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.088289 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3f8e-account-create-update-wcchj" event={"ID":"10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019","Type":"ContainerDied","Data":"2e672a1dded776e29e0bab7b213064d730fdd1545a03fbb5ffb80900119d6dd2"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.088352 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e672a1dded776e29e0bab7b213064d730fdd1545a03fbb5ffb80900119d6dd2" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.089149 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3f8e-account-create-update-wcchj" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.091238 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3a9a-account-create-update-wkmbs" event={"ID":"1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5","Type":"ContainerDied","Data":"e1a628004bc1dbabf6244b90b48f7b1987f7fca2663cda9faf62e2d61c9d2499"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.091285 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1a628004bc1dbabf6244b90b48f7b1987f7fca2663cda9faf62e2d61c9d2499" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.091373 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3a9a-account-create-update-wkmbs" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.097487 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e6bf-account-create-update-hhhqj" event={"ID":"b3946847-5d4a-4fcc-acac-7a49400db5b2","Type":"ContainerDied","Data":"a5acbfe665396b707daf03c8ed53e6cb10e5ea907d08f74e929fce54a19bed2c"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.097514 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5acbfe665396b707daf03c8ed53e6cb10e5ea907d08f74e929fce54a19bed2c" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.097591 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e6bf-account-create-update-hhhqj" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.101496 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8bl2c" event={"ID":"1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8","Type":"ContainerDied","Data":"a77006c8d435b1a1bf2d9a6f74127e5ad71393f5ebed72566b035fb4fc3677c5"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.101542 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a77006c8d435b1a1bf2d9a6f74127e5ad71393f5ebed72566b035fb4fc3677c5" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.101577 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8bl2c" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.105376 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4tzl\" (UniqueName: \"kubernetes.io/projected/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-kube-api-access-h4tzl\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.105424 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrfsz\" (UniqueName: \"kubernetes.io/projected/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-kube-api-access-nrfsz\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.105440 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.105453 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.114792 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tghwc" event={"ID":"5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7","Type":"ContainerDied","Data":"e4aa8760923ff895f626acfdfdd9d83423ac9ad87f6a854c00064b8698bd291c"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.114841 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4aa8760923ff895f626acfdfdd9d83423ac9ad87f6a854c00064b8698bd291c" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.114883 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tghwc" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.124396 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" event={"ID":"d2aace92-6ede-4554-9539-fd4f6bf4cfa6","Type":"ContainerStarted","Data":"f294b08ca2226f55b7ad3811c933d0e91f11e8da360d4b3e8b0a57135e9db880"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.140216 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5cvgg" event={"ID":"3607fd3f-e379-42ee-b778-5d1540371a35","Type":"ContainerDied","Data":"c75d542714faf7b86b82702669e39ba356f291f4698d8dd98f7c067d74bfd651"} Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.140268 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c75d542714faf7b86b82702669e39ba356f291f4698d8dd98f7c067d74bfd651" Dec 02 13:21:55 crc kubenswrapper[4725]: I1202 13:21:55.140350 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5cvgg" Dec 02 13:21:56 crc kubenswrapper[4725]: I1202 13:21:56.166027 4725 generic.go:334] "Generic (PLEG): container finished" podID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerID="d5081f4fe2afc34f7dfe0e3e6b66d91b62ae81f8fb97d28f93ff7676fb780338" exitCode=0 Dec 02 13:21:56 crc kubenswrapper[4725]: I1202 13:21:56.166087 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" event={"ID":"d2aace92-6ede-4554-9539-fd4f6bf4cfa6","Type":"ContainerDied","Data":"d5081f4fe2afc34f7dfe0e3e6b66d91b62ae81f8fb97d28f93ff7676fb780338"} Dec 02 13:21:59 crc kubenswrapper[4725]: I1202 13:21:59.192931 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zmwkf" event={"ID":"c76d8c86-54fe-4bb2-9590-76e2a56d8659","Type":"ContainerStarted","Data":"7937fc5aee77e9ecb0e9825cde9bfc317794afb497fb367d1c038e131d473544"} Dec 02 13:21:59 crc kubenswrapper[4725]: I1202 13:21:59.197106 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" event={"ID":"d2aace92-6ede-4554-9539-fd4f6bf4cfa6","Type":"ContainerStarted","Data":"571e8efba2a1c18cd0ecca1e18b0d5417ac5245d5e2d9b9108517b42b732778a"} Dec 02 13:21:59 crc kubenswrapper[4725]: I1202 13:21:59.197294 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:21:59 crc kubenswrapper[4725]: I1202 13:21:59.228638 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-zmwkf" podStartSLOduration=3.55307294 podStartE2EDuration="10.228618932s" podCreationTimestamp="2025-12-02 13:21:49 +0000 UTC" firstStartedPulling="2025-12-02 13:21:51.603826781 +0000 UTC m=+1042.560468476" lastFinishedPulling="2025-12-02 13:21:58.279372763 +0000 UTC m=+1049.236014468" observedRunningTime="2025-12-02 13:21:59.223073194 +0000 UTC m=+1050.179714889" watchObservedRunningTime="2025-12-02 13:21:59.228618932 +0000 UTC m=+1050.185260627" Dec 02 13:21:59 crc kubenswrapper[4725]: I1202 13:21:59.251915 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" podStartSLOduration=6.25189565 podStartE2EDuration="6.25189565s" podCreationTimestamp="2025-12-02 13:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:21:59.248003933 +0000 UTC m=+1050.204645638" watchObservedRunningTime="2025-12-02 13:21:59.25189565 +0000 UTC m=+1050.208537345" Dec 02 13:22:01 crc kubenswrapper[4725]: I1202 13:22:01.226206 4725 generic.go:334] "Generic (PLEG): container finished" podID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" containerID="6a99a7a44b1ca76d56407724b6adb597b9c0ff0392cc4225b7f17397752169c4" exitCode=0 Dec 02 13:22:01 crc kubenswrapper[4725]: I1202 13:22:01.226324 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7d27g" event={"ID":"154d76e0-4eac-4a29-8f3a-5aadb8762c6b","Type":"ContainerDied","Data":"6a99a7a44b1ca76d56407724b6adb597b9c0ff0392cc4225b7f17397752169c4"} Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.599578 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7d27g" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.655321 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lsp9\" (UniqueName: \"kubernetes.io/projected/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-kube-api-access-6lsp9\") pod \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.655373 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-combined-ca-bundle\") pod \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.655527 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-db-sync-config-data\") pod \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.655568 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-config-data\") pod \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\" (UID: \"154d76e0-4eac-4a29-8f3a-5aadb8762c6b\") " Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.662540 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-kube-api-access-6lsp9" (OuterVolumeSpecName: "kube-api-access-6lsp9") pod "154d76e0-4eac-4a29-8f3a-5aadb8762c6b" (UID: "154d76e0-4eac-4a29-8f3a-5aadb8762c6b"). InnerVolumeSpecName "kube-api-access-6lsp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.663644 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "154d76e0-4eac-4a29-8f3a-5aadb8762c6b" (UID: "154d76e0-4eac-4a29-8f3a-5aadb8762c6b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.682396 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "154d76e0-4eac-4a29-8f3a-5aadb8762c6b" (UID: "154d76e0-4eac-4a29-8f3a-5aadb8762c6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.713566 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-config-data" (OuterVolumeSpecName: "config-data") pod "154d76e0-4eac-4a29-8f3a-5aadb8762c6b" (UID: "154d76e0-4eac-4a29-8f3a-5aadb8762c6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.757157 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.757203 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lsp9\" (UniqueName: \"kubernetes.io/projected/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-kube-api-access-6lsp9\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.757220 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:02 crc kubenswrapper[4725]: I1202 13:22:02.757239 4725 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/154d76e0-4eac-4a29-8f3a-5aadb8762c6b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.242551 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7d27g" event={"ID":"154d76e0-4eac-4a29-8f3a-5aadb8762c6b","Type":"ContainerDied","Data":"7c794093c010bbce1a0ff33d881aa4912c76b9b9d4ce39567eed657ad817c475"} Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.242584 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c794093c010bbce1a0ff33d881aa4912c76b9b9d4ce39567eed657ad817c475" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.242633 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7d27g" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.244636 4725 generic.go:334] "Generic (PLEG): container finished" podID="c76d8c86-54fe-4bb2-9590-76e2a56d8659" containerID="7937fc5aee77e9ecb0e9825cde9bfc317794afb497fb367d1c038e131d473544" exitCode=0 Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.244678 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zmwkf" event={"ID":"c76d8c86-54fe-4bb2-9590-76e2a56d8659","Type":"ContainerDied","Data":"7937fc5aee77e9ecb0e9825cde9bfc317794afb497fb367d1c038e131d473544"} Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.693061 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fcphz"] Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.701222 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="dnsmasq-dns" containerID="cri-o://571e8efba2a1c18cd0ecca1e18b0d5417ac5245d5e2d9b9108517b42b732778a" gracePeriod=10 Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.708013 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739361 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-s6zhw"] Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739784 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3946847-5d4a-4fcc-acac-7a49400db5b2" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739804 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3946847-5d4a-4fcc-acac-7a49400db5b2" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739814 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" containerName="glance-db-sync" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739820 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" containerName="glance-db-sync" Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739829 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739836 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739853 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739861 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739869 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739875 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739891 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3607fd3f-e379-42ee-b778-5d1540371a35" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739897 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="3607fd3f-e379-42ee-b778-5d1540371a35" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: E1202 13:22:03.739916 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.739922 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740064 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="3607fd3f-e379-42ee-b778-5d1540371a35" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740074 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" containerName="glance-db-sync" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740089 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740098 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740108 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3946847-5d4a-4fcc-acac-7a49400db5b2" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740118 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" containerName="mariadb-database-create" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.740126 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" containerName="mariadb-account-create-update" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.743303 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.774090 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.774167 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbfrt\" (UniqueName: \"kubernetes.io/projected/8e291686-e6fe-49c4-b968-0dd3270b1ff5-kube-api-access-hbfrt\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.774220 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-config\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.774321 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.774349 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.774387 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.777000 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-s6zhw"] Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876103 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876144 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876176 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876214 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876243 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbfrt\" (UniqueName: \"kubernetes.io/projected/8e291686-e6fe-49c4-b968-0dd3270b1ff5-kube-api-access-hbfrt\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876274 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-config\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.876778 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.877178 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.877389 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.877548 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.877658 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-config\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.883389 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:03 crc kubenswrapper[4725]: I1202 13:22:03.903189 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbfrt\" (UniqueName: \"kubernetes.io/projected/8e291686-e6fe-49c4-b968-0dd3270b1ff5-kube-api-access-hbfrt\") pod \"dnsmasq-dns-74f6bcbc87-s6zhw\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.105849 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.255787 4725 generic.go:334] "Generic (PLEG): container finished" podID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerID="571e8efba2a1c18cd0ecca1e18b0d5417ac5245d5e2d9b9108517b42b732778a" exitCode=0 Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.255964 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" event={"ID":"d2aace92-6ede-4554-9539-fd4f6bf4cfa6","Type":"ContainerDied","Data":"571e8efba2a1c18cd0ecca1e18b0d5417ac5245d5e2d9b9108517b42b732778a"} Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.255988 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" event={"ID":"d2aace92-6ede-4554-9539-fd4f6bf4cfa6","Type":"ContainerDied","Data":"f294b08ca2226f55b7ad3811c933d0e91f11e8da360d4b3e8b0a57135e9db880"} Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.255998 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f294b08ca2226f55b7ad3811c933d0e91f11e8da360d4b3e8b0a57135e9db880" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.293440 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.398380 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-config\") pod \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.398550 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-svc\") pod \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.398866 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-swift-storage-0\") pod \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.398949 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-nb\") pod \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.398991 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz5dl\" (UniqueName: \"kubernetes.io/projected/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-kube-api-access-hz5dl\") pod \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.399024 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-sb\") pod \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\" (UID: \"d2aace92-6ede-4554-9539-fd4f6bf4cfa6\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.405061 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-kube-api-access-hz5dl" (OuterVolumeSpecName: "kube-api-access-hz5dl") pod "d2aace92-6ede-4554-9539-fd4f6bf4cfa6" (UID: "d2aace92-6ede-4554-9539-fd4f6bf4cfa6"). InnerVolumeSpecName "kube-api-access-hz5dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.441476 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-config" (OuterVolumeSpecName: "config") pod "d2aace92-6ede-4554-9539-fd4f6bf4cfa6" (UID: "d2aace92-6ede-4554-9539-fd4f6bf4cfa6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.450346 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d2aace92-6ede-4554-9539-fd4f6bf4cfa6" (UID: "d2aace92-6ede-4554-9539-fd4f6bf4cfa6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.461491 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d2aace92-6ede-4554-9539-fd4f6bf4cfa6" (UID: "d2aace92-6ede-4554-9539-fd4f6bf4cfa6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.476821 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d2aace92-6ede-4554-9539-fd4f6bf4cfa6" (UID: "d2aace92-6ede-4554-9539-fd4f6bf4cfa6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.477096 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d2aace92-6ede-4554-9539-fd4f6bf4cfa6" (UID: "d2aace92-6ede-4554-9539-fd4f6bf4cfa6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.500842 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz5dl\" (UniqueName: \"kubernetes.io/projected/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-kube-api-access-hz5dl\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.500880 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.500892 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.500908 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.500918 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.500928 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d2aace92-6ede-4554-9539-fd4f6bf4cfa6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.579806 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.634971 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-s6zhw"] Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.704547 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-combined-ca-bundle\") pod \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.705338 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fhjd\" (UniqueName: \"kubernetes.io/projected/c76d8c86-54fe-4bb2-9590-76e2a56d8659-kube-api-access-7fhjd\") pod \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.705711 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-config-data\") pod \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\" (UID: \"c76d8c86-54fe-4bb2-9590-76e2a56d8659\") " Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.711918 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c76d8c86-54fe-4bb2-9590-76e2a56d8659-kube-api-access-7fhjd" (OuterVolumeSpecName: "kube-api-access-7fhjd") pod "c76d8c86-54fe-4bb2-9590-76e2a56d8659" (UID: "c76d8c86-54fe-4bb2-9590-76e2a56d8659"). InnerVolumeSpecName "kube-api-access-7fhjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.728291 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c76d8c86-54fe-4bb2-9590-76e2a56d8659" (UID: "c76d8c86-54fe-4bb2-9590-76e2a56d8659"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.763444 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-config-data" (OuterVolumeSpecName: "config-data") pod "c76d8c86-54fe-4bb2-9590-76e2a56d8659" (UID: "c76d8c86-54fe-4bb2-9590-76e2a56d8659"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.809018 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.809055 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fhjd\" (UniqueName: \"kubernetes.io/projected/c76d8c86-54fe-4bb2-9590-76e2a56d8659-kube-api-access-7fhjd\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:04 crc kubenswrapper[4725]: I1202 13:22:04.809068 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c76d8c86-54fe-4bb2-9590-76e2a56d8659-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.264130 4725 generic.go:334] "Generic (PLEG): container finished" podID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerID="1631ee1a0d7649d2eb325b3b5d861323a5276e3e002133eb9afab2ed5d013fd3" exitCode=0 Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.264196 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" event={"ID":"8e291686-e6fe-49c4-b968-0dd3270b1ff5","Type":"ContainerDied","Data":"1631ee1a0d7649d2eb325b3b5d861323a5276e3e002133eb9afab2ed5d013fd3"} Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.264243 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" event={"ID":"8e291686-e6fe-49c4-b968-0dd3270b1ff5","Type":"ContainerStarted","Data":"da504358b157c02dc97fc129efa689d727e544216b01d6092f0741729ff92404"} Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.265876 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zmwkf" event={"ID":"c76d8c86-54fe-4bb2-9590-76e2a56d8659","Type":"ContainerDied","Data":"f417cc723ab481c39589937b86dcc63bfed7d8d852443198a6135c591f79ac30"} Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.265900 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zmwkf" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.265913 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f417cc723ab481c39589937b86dcc63bfed7d8d852443198a6135c591f79ac30" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.265900 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-fcphz" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.515787 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fcphz"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.534584 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-fcphz"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.566583 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-s6zhw"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.583248 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b6q6j"] Dec 02 13:22:05 crc kubenswrapper[4725]: E1202 13:22:05.583666 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76d8c86-54fe-4bb2-9590-76e2a56d8659" containerName="keystone-db-sync" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.583681 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76d8c86-54fe-4bb2-9590-76e2a56d8659" containerName="keystone-db-sync" Dec 02 13:22:05 crc kubenswrapper[4725]: E1202 13:22:05.583704 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="init" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.583710 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="init" Dec 02 13:22:05 crc kubenswrapper[4725]: E1202 13:22:05.583726 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="dnsmasq-dns" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.583732 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="dnsmasq-dns" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.583898 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="c76d8c86-54fe-4bb2-9590-76e2a56d8659" containerName="keystone-db-sync" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.583924 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" containerName="dnsmasq-dns" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.584552 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.590826 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-twszr" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.591082 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.591192 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.591305 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.591412 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.610390 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b6q6j"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.630408 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-65wqq"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.632375 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.648336 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-scripts\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.648406 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-fernet-keys\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.648438 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6rw5\" (UniqueName: \"kubernetes.io/projected/649811b5-8aff-4f44-8799-850000589dbf-kube-api-access-c6rw5\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.648521 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-combined-ca-bundle\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.648592 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-credential-keys\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.648613 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-config-data\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.657052 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-65wqq"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751435 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751535 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-credential-keys\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751565 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-config-data\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751593 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwsb\" (UniqueName: \"kubernetes.io/projected/14a594f2-c624-4fde-85d2-c17cb5501c51-kube-api-access-6cwsb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751620 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-scripts\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751681 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-svc\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751711 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751741 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-fernet-keys\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751774 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6rw5\" (UniqueName: \"kubernetes.io/projected/649811b5-8aff-4f44-8799-850000589dbf-kube-api-access-c6rw5\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751854 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-combined-ca-bundle\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751879 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-config\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.751936 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.763070 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-fernet-keys\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.771231 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-scripts\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.773048 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-credential-keys\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.774158 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-config-data\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.775560 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-combined-ca-bundle\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.781014 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6rw5\" (UniqueName: \"kubernetes.io/projected/649811b5-8aff-4f44-8799-850000589dbf-kube-api-access-c6rw5\") pod \"keystone-bootstrap-b6q6j\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.814538 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5489f8c44c-g22fg"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.816586 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.825365 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5489f8c44c-g22fg"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.835013 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.835373 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-nc2hs" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.835562 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.835663 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.864851 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-lx9dr"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.865244 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-config\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.865348 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.865395 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.865499 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwsb\" (UniqueName: \"kubernetes.io/projected/14a594f2-c624-4fde-85d2-c17cb5501c51-kube-api-access-6cwsb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.865566 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-svc\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.865596 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.866009 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.866674 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.867376 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.867780 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-config\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.868128 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-svc\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.873028 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.876173 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.876365 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wmgxq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.876486 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.921922 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lx9dr"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.925871 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.943602 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwsb\" (UniqueName: \"kubernetes.io/projected/14a594f2-c624-4fde-85d2-c17cb5501c51-kube-api-access-6cwsb\") pod \"dnsmasq-dns-847c4cc679-65wqq\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966704 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-scripts\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966759 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-config\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966782 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de526deb-a82c-469a-9387-308555f0d667-logs\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966816 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mppxv\" (UniqueName: \"kubernetes.io/projected/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-kube-api-access-mppxv\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966834 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-config-data\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966863 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9zf7\" (UniqueName: \"kubernetes.io/projected/de526deb-a82c-469a-9387-308555f0d667-kube-api-access-m9zf7\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966919 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/de526deb-a82c-469a-9387-308555f0d667-horizon-secret-key\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.966946 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-combined-ca-bundle\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.967524 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-pdwcp"] Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.968951 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.969833 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.972448 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zd7gg" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.973230 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 02 13:22:05 crc kubenswrapper[4725]: I1202 13:22:05.974483 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.013246 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pdwcp"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068289 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-db-sync-config-data\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068340 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-combined-ca-bundle\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068364 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mppxv\" (UniqueName: \"kubernetes.io/projected/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-kube-api-access-mppxv\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068386 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-config-data\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068403 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9zf7\" (UniqueName: \"kubernetes.io/projected/de526deb-a82c-469a-9387-308555f0d667-kube-api-access-m9zf7\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068420 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-config-data\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068439 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11048035-cfbd-4e03-8427-83e0d2a91a63-etc-machine-id\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068493 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/de526deb-a82c-469a-9387-308555f0d667-horizon-secret-key\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068520 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-combined-ca-bundle\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068548 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96b4p\" (UniqueName: \"kubernetes.io/projected/11048035-cfbd-4e03-8427-83e0d2a91a63-kube-api-access-96b4p\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068566 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-scripts\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068607 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-scripts\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068643 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-config\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.068659 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de526deb-a82c-469a-9387-308555f0d667-logs\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.069051 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de526deb-a82c-469a-9387-308555f0d667-logs\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.070272 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-config-data\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.078136 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-scripts\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.085536 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/de526deb-a82c-469a-9387-308555f0d667-horizon-secret-key\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.087510 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-combined-ca-bundle\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.097277 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-config\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.103289 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9zf7\" (UniqueName: \"kubernetes.io/projected/de526deb-a82c-469a-9387-308555f0d667-kube-api-access-m9zf7\") pod \"horizon-5489f8c44c-g22fg\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.117966 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mppxv\" (UniqueName: \"kubernetes.io/projected/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-kube-api-access-mppxv\") pod \"neutron-db-sync-lx9dr\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.134523 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-9blbh"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.135704 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.137867 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6h2cf" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.138285 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.183410 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-combined-ca-bundle\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.183480 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-config-data\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.183503 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11048035-cfbd-4e03-8427-83e0d2a91a63-etc-machine-id\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.183560 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96b4p\" (UniqueName: \"kubernetes.io/projected/11048035-cfbd-4e03-8427-83e0d2a91a63-kube-api-access-96b4p\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.183574 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-scripts\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.183657 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-db-sync-config-data\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.184874 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11048035-cfbd-4e03-8427-83e0d2a91a63-etc-machine-id\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.190009 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.193151 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-db-sync-config-data\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.193226 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9blbh"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.193552 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-combined-ca-bundle\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.199056 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.230024 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-scripts\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.230852 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-config-data\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.267219 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96b4p\" (UniqueName: \"kubernetes.io/projected/11048035-cfbd-4e03-8427-83e0d2a91a63-kube-api-access-96b4p\") pod \"cinder-db-sync-pdwcp\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.295617 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.297170 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9nbc\" (UniqueName: \"kubernetes.io/projected/52677bda-0bda-4d69-af01-f6ce198582f6-kube-api-access-z9nbc\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.310237 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-db-sync-config-data\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.310587 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.310596 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-combined-ca-bundle\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.313287 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" event={"ID":"8e291686-e6fe-49c4-b968-0dd3270b1ff5","Type":"ContainerStarted","Data":"927b30d9658bf3cbfa19c8628e5515c36fd987b1b94abe0e79ed31d47a829d31"} Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.319053 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.316130 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.319928 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.320802 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.330737 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-65wqq"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.347589 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-psww9"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.349126 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-psww9" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.357797 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-psww9"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.359350 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-jqqg2" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.359720 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.359944 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.386804 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74f4bf4679-pfzbz"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.388377 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.415569 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417603 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9nbc\" (UniqueName: \"kubernetes.io/projected/52677bda-0bda-4d69-af01-f6ce198582f6-kube-api-access-z9nbc\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417696 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-db-sync-config-data\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417742 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417823 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5r6w\" (UniqueName: \"kubernetes.io/projected/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-kube-api-access-g5r6w\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417868 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-config-data\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417896 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-combined-ca-bundle\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417930 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-run-httpd\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.417968 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-log-httpd\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.418075 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-scripts\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.418189 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.434132 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-combined-ca-bundle\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.436917 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-db-sync-config-data\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.441343 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.454656 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9nbc\" (UniqueName: \"kubernetes.io/projected/52677bda-0bda-4d69-af01-f6ce198582f6-kube-api-access-z9nbc\") pod \"barbican-db-sync-9blbh\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.455077 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.456558 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.463009 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-c5cmk" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.497492 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.500676 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74f4bf4679-pfzbz"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533709 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533769 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-logs\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533794 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-config-data\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533826 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533849 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533880 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-combined-ca-bundle\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533919 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgxld\" (UniqueName: \"kubernetes.io/projected/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-kube-api-access-lgxld\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533947 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-logs\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533967 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.533986 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534022 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-scripts\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534053 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df2ed35d-e2f2-4c5f-b2db-12d95291d167-logs\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534071 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5r6w\" (UniqueName: \"kubernetes.io/projected/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-kube-api-access-g5r6w\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534086 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-config-data\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534101 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-scripts\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534121 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rtmv\" (UniqueName: \"kubernetes.io/projected/df2ed35d-e2f2-4c5f-b2db-12d95291d167-kube-api-access-9rtmv\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534141 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-config-data\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534171 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-run-httpd\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534192 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df2ed35d-e2f2-4c5f-b2db-12d95291d167-horizon-secret-key\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534219 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-log-httpd\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534260 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-scripts\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534286 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z656k\" (UniqueName: \"kubernetes.io/projected/e2bbca85-b50a-46b4-857e-6e3a2217b5da-kube-api-access-z656k\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534315 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-scripts\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.534330 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-config-data\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.537413 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-log-httpd\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.537596 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fxcbw"] Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.539677 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.542423 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-run-httpd\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.543250 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-config-data\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.543372 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.548874 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:06 crc kubenswrapper[4725]: I1202 13:22:06.549138 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9blbh" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.556983 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-scripts\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.583920 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5r6w\" (UniqueName: \"kubernetes.io/projected/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-kube-api-access-g5r6w\") pod \"ceilometer-0\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " pod="openstack/ceilometer-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.583973 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635690 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkfvn\" (UniqueName: \"kubernetes.io/projected/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-kube-api-access-pkfvn\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635736 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635764 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-logs\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635783 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-config-data\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635824 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635852 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-combined-ca-bundle\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635874 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgxld\" (UniqueName: \"kubernetes.io/projected/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-kube-api-access-lgxld\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635910 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635932 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-logs\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.635962 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636004 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-scripts\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636027 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df2ed35d-e2f2-4c5f-b2db-12d95291d167-logs\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636046 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-config-data\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636068 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-scripts\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636092 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rtmv\" (UniqueName: \"kubernetes.io/projected/df2ed35d-e2f2-4c5f-b2db-12d95291d167-kube-api-access-9rtmv\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636113 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636143 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df2ed35d-e2f2-4c5f-b2db-12d95291d167-horizon-secret-key\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636170 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636202 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z656k\" (UniqueName: \"kubernetes.io/projected/e2bbca85-b50a-46b4-857e-6e3a2217b5da-kube-api-access-z656k\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636227 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-scripts\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636275 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-config-data\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.636295 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-config\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.637147 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-scripts\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.637424 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df2ed35d-e2f2-4c5f-b2db-12d95291d167-logs\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.637913 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.640327 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.640538 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.643073 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-config-data\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.643482 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-combined-ca-bundle\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.643982 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-logs\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.644203 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-logs\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.646901 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.648046 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df2ed35d-e2f2-4c5f-b2db-12d95291d167-horizon-secret-key\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.649111 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-config-data\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.649361 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-scripts\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.655896 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-config-data\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.659350 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-scripts\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.661646 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgxld\" (UniqueName: \"kubernetes.io/projected/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-kube-api-access-lgxld\") pod \"placement-db-sync-psww9\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.665047 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z656k\" (UniqueName: \"kubernetes.io/projected/e2bbca85-b50a-46b4-857e-6e3a2217b5da-kube-api-access-z656k\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.668689 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fxcbw"] Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.670442 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rtmv\" (UniqueName: \"kubernetes.io/projected/df2ed35d-e2f2-4c5f-b2db-12d95291d167-kube-api-access-9rtmv\") pod \"horizon-74f4bf4679-pfzbz\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.673857 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.676440 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.678871 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.682572 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.684203 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.694186 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-psww9" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.701801 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" podStartSLOduration=3.701779957 podStartE2EDuration="3.701779957s" podCreationTimestamp="2025-12-02 13:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:06.415244113 +0000 UTC m=+1057.371885808" watchObservedRunningTime="2025-12-02 13:22:06.701779957 +0000 UTC m=+1057.658421652" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737609 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-logs\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737692 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737742 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737800 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737837 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737871 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737908 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737935 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737958 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.737983 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.738008 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-config\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.738038 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdqhk\" (UniqueName: \"kubernetes.io/projected/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-kube-api-access-sdqhk\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.738063 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkfvn\" (UniqueName: \"kubernetes.io/projected/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-kube-api-access-pkfvn\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.739042 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.739317 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.739770 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.740076 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-config\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.740332 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.740425 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.757050 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkfvn\" (UniqueName: \"kubernetes.io/projected/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-kube-api-access-pkfvn\") pod \"dnsmasq-dns-785d8bcb8c-fxcbw\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.807837 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.853819 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.853916 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.853970 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.853992 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.854015 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.854042 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdqhk\" (UniqueName: \"kubernetes.io/projected/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-kube-api-access-sdqhk\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.854099 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-logs\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.854921 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-logs\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.857025 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.859329 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.861916 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.872826 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.873786 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.885595 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.894327 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdqhk\" (UniqueName: \"kubernetes.io/projected/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-kube-api-access-sdqhk\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.908880 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:06.933474 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b6q6j"] Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:07.082855 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-65wqq"] Dec 02 13:22:07 crc kubenswrapper[4725]: W1202 13:22:07.213235 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14a594f2_c624_4fde_85d2_c17cb5501c51.slice/crio-f083c4a65d4fe3681a694e81f909b0962831cc67e248eb53016c48d14d2cb1b6 WatchSource:0}: Error finding container f083c4a65d4fe3681a694e81f909b0962831cc67e248eb53016c48d14d2cb1b6: Status 404 returned error can't find the container with id f083c4a65d4fe3681a694e81f909b0962831cc67e248eb53016c48d14d2cb1b6 Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:07.278971 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2aace92-6ede-4554-9539-fd4f6bf4cfa6" path="/var/lib/kubelet/pods/d2aace92-6ede-4554-9539-fd4f6bf4cfa6/volumes" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:07.282274 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:07.322526 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b6q6j" event={"ID":"649811b5-8aff-4f44-8799-850000589dbf","Type":"ContainerStarted","Data":"ced5c87f870a9ae316ec06cae78dd0c8bbe5437cd49d50a8008f37245edcb44c"} Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:07.324234 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" event={"ID":"14a594f2-c624-4fde-85d2-c17cb5501c51","Type":"ContainerStarted","Data":"f083c4a65d4fe3681a694e81f909b0962831cc67e248eb53016c48d14d2cb1b6"} Dec 02 13:22:07 crc kubenswrapper[4725]: I1202 13:22:07.324487 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerName="dnsmasq-dns" containerID="cri-o://927b30d9658bf3cbfa19c8628e5515c36fd987b1b94abe0e79ed31d47a829d31" gracePeriod=10 Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.245160 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5489f8c44c-g22fg"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.298281 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lx9dr"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.372078 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lx9dr" event={"ID":"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76","Type":"ContainerStarted","Data":"34e6b07e25c1fb2c0b4892de6b2ce5f2ee4937293303949b761bd9752cd24c56"} Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.374701 4725 generic.go:334] "Generic (PLEG): container finished" podID="14a594f2-c624-4fde-85d2-c17cb5501c51" containerID="e12d6f342f00a4e8cd4ce9a77354bbf9a738b349f060f6b2ac2387eb38d74f5c" exitCode=0 Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.374745 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" event={"ID":"14a594f2-c624-4fde-85d2-c17cb5501c51","Type":"ContainerDied","Data":"e12d6f342f00a4e8cd4ce9a77354bbf9a738b349f060f6b2ac2387eb38d74f5c"} Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.377191 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b6q6j" event={"ID":"649811b5-8aff-4f44-8799-850000589dbf","Type":"ContainerStarted","Data":"fd8145ab7dcf6d7a1d56c5b97a382302a90ddece38539fefd1b7a4258e63a35a"} Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.389023 4725 generic.go:334] "Generic (PLEG): container finished" podID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerID="927b30d9658bf3cbfa19c8628e5515c36fd987b1b94abe0e79ed31d47a829d31" exitCode=0 Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.389371 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" event={"ID":"8e291686-e6fe-49c4-b968-0dd3270b1ff5","Type":"ContainerDied","Data":"927b30d9658bf3cbfa19c8628e5515c36fd987b1b94abe0e79ed31d47a829d31"} Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.392888 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5489f8c44c-g22fg" event={"ID":"de526deb-a82c-469a-9387-308555f0d667","Type":"ContainerStarted","Data":"064d9a388db5e9c5434402b57a424de1e6fde35360bb319507863b9c5bafc9eb"} Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.424664 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b6q6j" podStartSLOduration=3.424645702 podStartE2EDuration="3.424645702s" podCreationTimestamp="2025-12-02 13:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:08.421636808 +0000 UTC m=+1059.378278503" watchObservedRunningTime="2025-12-02 13:22:08.424645702 +0000 UTC m=+1059.381287397" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.729502 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.815482 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830690 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbfrt\" (UniqueName: \"kubernetes.io/projected/8e291686-e6fe-49c4-b968-0dd3270b1ff5-kube-api-access-hbfrt\") pod \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830756 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cwsb\" (UniqueName: \"kubernetes.io/projected/14a594f2-c624-4fde-85d2-c17cb5501c51-kube-api-access-6cwsb\") pod \"14a594f2-c624-4fde-85d2-c17cb5501c51\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830798 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-svc\") pod \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830835 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-config\") pod \"14a594f2-c624-4fde-85d2-c17cb5501c51\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830878 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-nb\") pod \"14a594f2-c624-4fde-85d2-c17cb5501c51\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830934 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-sb\") pod \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.830961 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-config\") pod \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.831038 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-swift-storage-0\") pod \"14a594f2-c624-4fde-85d2-c17cb5501c51\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.831070 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-svc\") pod \"14a594f2-c624-4fde-85d2-c17cb5501c51\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.831104 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-swift-storage-0\") pod \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.831215 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-nb\") pod \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\" (UID: \"8e291686-e6fe-49c4-b968-0dd3270b1ff5\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.831234 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-sb\") pod \"14a594f2-c624-4fde-85d2-c17cb5501c51\" (UID: \"14a594f2-c624-4fde-85d2-c17cb5501c51\") " Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.853765 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e291686-e6fe-49c4-b968-0dd3270b1ff5-kube-api-access-hbfrt" (OuterVolumeSpecName: "kube-api-access-hbfrt") pod "8e291686-e6fe-49c4-b968-0dd3270b1ff5" (UID: "8e291686-e6fe-49c4-b968-0dd3270b1ff5"). InnerVolumeSpecName "kube-api-access-hbfrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.865349 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a594f2-c624-4fde-85d2-c17cb5501c51-kube-api-access-6cwsb" (OuterVolumeSpecName: "kube-api-access-6cwsb") pod "14a594f2-c624-4fde-85d2-c17cb5501c51" (UID: "14a594f2-c624-4fde-85d2-c17cb5501c51"). InnerVolumeSpecName "kube-api-access-6cwsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.886202 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-psww9"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.908670 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.916641 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e291686-e6fe-49c4-b968-0dd3270b1ff5" (UID: "8e291686-e6fe-49c4-b968-0dd3270b1ff5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.933947 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.934375 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbfrt\" (UniqueName: \"kubernetes.io/projected/8e291686-e6fe-49c4-b968-0dd3270b1ff5-kube-api-access-hbfrt\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.934396 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cwsb\" (UniqueName: \"kubernetes.io/projected/14a594f2-c624-4fde-85d2-c17cb5501c51-kube-api-access-6cwsb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.934865 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14a594f2-c624-4fde-85d2-c17cb5501c51" (UID: "14a594f2-c624-4fde-85d2-c17cb5501c51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.950283 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-9blbh"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.969975 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-pdwcp"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.974222 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14a594f2-c624-4fde-85d2-c17cb5501c51" (UID: "14a594f2-c624-4fde-85d2-c17cb5501c51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.976162 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74f4bf4679-pfzbz"] Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.983420 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fxcbw"] Dec 02 13:22:08 crc kubenswrapper[4725]: W1202 13:22:08.989443 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52677bda_0bda_4d69_af01_f6ce198582f6.slice/crio-1050fbb08b2302a2f38baf6c4b201214cf518b366ce2bddc683cbe2e981baf54 WatchSource:0}: Error finding container 1050fbb08b2302a2f38baf6c4b201214cf518b366ce2bddc683cbe2e981baf54: Status 404 returned error can't find the container with id 1050fbb08b2302a2f38baf6c4b201214cf518b366ce2bddc683cbe2e981baf54 Dec 02 13:22:08 crc kubenswrapper[4725]: W1202 13:22:08.991917 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c262e77_097d_4bd1_aebf_dc0a401fdb5b.slice/crio-29e0cbdb8a7e70bdecde52a551306a6ac1f778f2340130f605b704cd3e66d804 WatchSource:0}: Error finding container 29e0cbdb8a7e70bdecde52a551306a6ac1f778f2340130f605b704cd3e66d804: Status 404 returned error can't find the container with id 29e0cbdb8a7e70bdecde52a551306a6ac1f778f2340130f605b704cd3e66d804 Dec 02 13:22:08 crc kubenswrapper[4725]: I1202 13:22:08.996485 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e291686-e6fe-49c4-b968-0dd3270b1ff5" (UID: "8e291686-e6fe-49c4-b968-0dd3270b1ff5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.036479 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.036622 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.036634 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.043956 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14a594f2-c624-4fde-85d2-c17cb5501c51" (UID: "14a594f2-c624-4fde-85d2-c17cb5501c51"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.051464 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e291686-e6fe-49c4-b968-0dd3270b1ff5" (UID: "8e291686-e6fe-49c4-b968-0dd3270b1ff5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.063180 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14a594f2-c624-4fde-85d2-c17cb5501c51" (UID: "14a594f2-c624-4fde-85d2-c17cb5501c51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.069270 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.077176 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-config" (OuterVolumeSpecName: "config") pod "14a594f2-c624-4fde-85d2-c17cb5501c51" (UID: "14a594f2-c624-4fde-85d2-c17cb5501c51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.083205 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e291686-e6fe-49c4-b968-0dd3270b1ff5" (UID: "8e291686-e6fe-49c4-b968-0dd3270b1ff5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.086759 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-config" (OuterVolumeSpecName: "config") pod "8e291686-e6fe-49c4-b968-0dd3270b1ff5" (UID: "8e291686-e6fe-49c4-b968-0dd3270b1ff5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.138920 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.138954 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.138966 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.138977 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.138989 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e291686-e6fe-49c4-b968-0dd3270b1ff5-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.139001 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14a594f2-c624-4fde-85d2-c17cb5501c51-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.415728 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9blbh" event={"ID":"52677bda-0bda-4d69-af01-f6ce198582f6","Type":"ContainerStarted","Data":"1050fbb08b2302a2f38baf6c4b201214cf518b366ce2bddc683cbe2e981baf54"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.427709 4725 generic.go:334] "Generic (PLEG): container finished" podID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerID="b0bb34439059e7590dcd8a6433c94ed8a900084ea5458c0fe445eb78b85f0c99" exitCode=0 Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.427824 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" event={"ID":"9c262e77-097d-4bd1-aebf-dc0a401fdb5b","Type":"ContainerDied","Data":"b0bb34439059e7590dcd8a6433c94ed8a900084ea5458c0fe445eb78b85f0c99"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.427857 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" event={"ID":"9c262e77-097d-4bd1-aebf-dc0a401fdb5b","Type":"ContainerStarted","Data":"29e0cbdb8a7e70bdecde52a551306a6ac1f778f2340130f605b704cd3e66d804"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.435264 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerStarted","Data":"98355c053203c3bcb96869dbcbccd637a1baf8b8bc9356eb1ab99898b44abe87"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.439536 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lx9dr" event={"ID":"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76","Type":"ContainerStarted","Data":"39ea52ea6285c01b3b519293f937ef8f9d19b4e72bc46fa935f7e64f87e1b08d"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.446452 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdwcp" event={"ID":"11048035-cfbd-4e03-8427-83e0d2a91a63","Type":"ContainerStarted","Data":"e908e15b5237bd932306de557c2e39a553aa6f06c0b7c9afc7badd405b6bea8d"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.457948 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.458109 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-65wqq" event={"ID":"14a594f2-c624-4fde-85d2-c17cb5501c51","Type":"ContainerDied","Data":"f083c4a65d4fe3681a694e81f909b0962831cc67e248eb53016c48d14d2cb1b6"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.458250 4725 scope.go:117] "RemoveContainer" containerID="e12d6f342f00a4e8cd4ce9a77354bbf9a738b349f060f6b2ac2387eb38d74f5c" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.466977 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2bbca85-b50a-46b4-857e-6e3a2217b5da","Type":"ContainerStarted","Data":"953369eb2daef17e909e219a0feae590bc4a1db6a6560c33bb98608b1b2468d2"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.471679 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" event={"ID":"8e291686-e6fe-49c4-b968-0dd3270b1ff5","Type":"ContainerDied","Data":"da504358b157c02dc97fc129efa689d727e544216b01d6092f0741729ff92404"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.471935 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-s6zhw" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.483558 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f4bf4679-pfzbz" event={"ID":"df2ed35d-e2f2-4c5f-b2db-12d95291d167","Type":"ContainerStarted","Data":"8fcc5da26dcda3cd55b724aef41257433aae3101c0020fceb510f16c12a8c2f9"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.486320 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-psww9" event={"ID":"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e","Type":"ContainerStarted","Data":"7eecbf2783c4d01dea1be3a0b5725341ce4a1ab1d5fd14e66878d24112b90e65"} Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.502620 4725 scope.go:117] "RemoveContainer" containerID="927b30d9658bf3cbfa19c8628e5515c36fd987b1b94abe0e79ed31d47a829d31" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.595008 4725 scope.go:117] "RemoveContainer" containerID="1631ee1a0d7649d2eb325b3b5d861323a5276e3e002133eb9afab2ed5d013fd3" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.608885 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-s6zhw"] Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.660621 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-s6zhw"] Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.681329 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-65wqq"] Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.701180 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-65wqq"] Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.728345 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-lx9dr" podStartSLOduration=4.728323101 podStartE2EDuration="4.728323101s" podCreationTimestamp="2025-12-02 13:22:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:09.655703808 +0000 UTC m=+1060.612345503" watchObservedRunningTime="2025-12-02 13:22:09.728323101 +0000 UTC m=+1060.684964796" Dec 02 13:22:09 crc kubenswrapper[4725]: I1202 13:22:09.822485 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:09 crc kubenswrapper[4725]: W1202 13:22:09.837483 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99b94af9_f4c3_4acd_bb4c_50d54ecaae4a.slice/crio-77c84d5f2003f710e37cbe8807629d297489eb10373ddefc1a7e63719b644d7c WatchSource:0}: Error finding container 77c84d5f2003f710e37cbe8807629d297489eb10373ddefc1a7e63719b644d7c: Status 404 returned error can't find the container with id 77c84d5f2003f710e37cbe8807629d297489eb10373ddefc1a7e63719b644d7c Dec 02 13:22:10 crc kubenswrapper[4725]: I1202 13:22:10.572719 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2bbca85-b50a-46b4-857e-6e3a2217b5da","Type":"ContainerStarted","Data":"3b7f664409d34c4e26d27f826855e5d9ee8a05571c9af802547b8c790b55c94d"} Dec 02 13:22:10 crc kubenswrapper[4725]: I1202 13:22:10.575045 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" event={"ID":"9c262e77-097d-4bd1-aebf-dc0a401fdb5b","Type":"ContainerStarted","Data":"3d89219dd82c7d2ce15dd379b4c86cf34afed0f7171061fa2878b6089deb26fa"} Dec 02 13:22:10 crc kubenswrapper[4725]: I1202 13:22:10.576230 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:10 crc kubenswrapper[4725]: I1202 13:22:10.578984 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a","Type":"ContainerStarted","Data":"77c84d5f2003f710e37cbe8807629d297489eb10373ddefc1a7e63719b644d7c"} Dec 02 13:22:10 crc kubenswrapper[4725]: I1202 13:22:10.636428 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" podStartSLOduration=4.636407027 podStartE2EDuration="4.636407027s" podCreationTimestamp="2025-12-02 13:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:10.622277796 +0000 UTC m=+1061.578919491" watchObservedRunningTime="2025-12-02 13:22:10.636407027 +0000 UTC m=+1061.593048732" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.083367 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.171568 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.229893 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5489f8c44c-g22fg"] Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.237927 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.281583 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a594f2-c624-4fde-85d2-c17cb5501c51" path="/var/lib/kubelet/pods/14a594f2-c624-4fde-85d2-c17cb5501c51/volumes" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.282112 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" path="/var/lib/kubelet/pods/8e291686-e6fe-49c4-b968-0dd3270b1ff5/volumes" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.288036 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77f7bfffdc-cnbqk"] Dec 02 13:22:11 crc kubenswrapper[4725]: E1202 13:22:11.288526 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a594f2-c624-4fde-85d2-c17cb5501c51" containerName="init" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.288568 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a594f2-c624-4fde-85d2-c17cb5501c51" containerName="init" Dec 02 13:22:11 crc kubenswrapper[4725]: E1202 13:22:11.288601 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerName="dnsmasq-dns" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.288610 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerName="dnsmasq-dns" Dec 02 13:22:11 crc kubenswrapper[4725]: E1202 13:22:11.288623 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerName="init" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.288632 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerName="init" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.288866 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e291686-e6fe-49c4-b968-0dd3270b1ff5" containerName="dnsmasq-dns" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.288887 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a594f2-c624-4fde-85d2-c17cb5501c51" containerName="init" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.309450 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77f7bfffdc-cnbqk"] Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.309576 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.386726 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-scripts\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.387051 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n46wz\" (UniqueName: \"kubernetes.io/projected/5352a37f-00de-4087-8337-6536a0c03fd5-kube-api-access-n46wz\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.387444 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5352a37f-00de-4087-8337-6536a0c03fd5-logs\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.387555 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-config-data\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.387580 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5352a37f-00de-4087-8337-6536a0c03fd5-horizon-secret-key\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.489079 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n46wz\" (UniqueName: \"kubernetes.io/projected/5352a37f-00de-4087-8337-6536a0c03fd5-kube-api-access-n46wz\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.489117 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-scripts\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.489253 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5352a37f-00de-4087-8337-6536a0c03fd5-logs\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.489277 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-config-data\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.489342 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5352a37f-00de-4087-8337-6536a0c03fd5-horizon-secret-key\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.489645 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5352a37f-00de-4087-8337-6536a0c03fd5-logs\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.490269 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-scripts\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.490529 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-config-data\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.496675 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5352a37f-00de-4087-8337-6536a0c03fd5-horizon-secret-key\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.507827 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n46wz\" (UniqueName: \"kubernetes.io/projected/5352a37f-00de-4087-8337-6536a0c03fd5-kube-api-access-n46wz\") pod \"horizon-77f7bfffdc-cnbqk\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.606287 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a","Type":"ContainerStarted","Data":"0cfc71d4d21cd4c9402885507cfe656bf22e7953c00eb938c1887acdcf68cd81"} Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.636591 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2bbca85-b50a-46b4-857e-6e3a2217b5da","Type":"ContainerStarted","Data":"50f3d76cd46ff7389f603110bce47913ecba235795a58bff55804b656e6c4e01"} Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.636644 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-httpd" containerID="cri-o://50f3d76cd46ff7389f603110bce47913ecba235795a58bff55804b656e6c4e01" gracePeriod=30 Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.636313 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-log" containerID="cri-o://3b7f664409d34c4e26d27f826855e5d9ee8a05571c9af802547b8c790b55c94d" gracePeriod=30 Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.644126 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:11 crc kubenswrapper[4725]: I1202 13:22:11.670910 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.670884952 podStartE2EDuration="5.670884952s" podCreationTimestamp="2025-12-02 13:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:11.663207101 +0000 UTC m=+1062.619848796" watchObservedRunningTime="2025-12-02 13:22:11.670884952 +0000 UTC m=+1062.627526647" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.360297 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77f7bfffdc-cnbqk"] Dec 02 13:22:12 crc kubenswrapper[4725]: W1202 13:22:12.399725 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5352a37f_00de_4087_8337_6536a0c03fd5.slice/crio-9d961ffdbca1348bf88e054d0fe9890b289b17b4edb6a1b107a5f4aadb948695 WatchSource:0}: Error finding container 9d961ffdbca1348bf88e054d0fe9890b289b17b4edb6a1b107a5f4aadb948695: Status 404 returned error can't find the container with id 9d961ffdbca1348bf88e054d0fe9890b289b17b4edb6a1b107a5f4aadb948695 Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.651061 4725 generic.go:334] "Generic (PLEG): container finished" podID="649811b5-8aff-4f44-8799-850000589dbf" containerID="fd8145ab7dcf6d7a1d56c5b97a382302a90ddece38539fefd1b7a4258e63a35a" exitCode=0 Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.651091 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b6q6j" event={"ID":"649811b5-8aff-4f44-8799-850000589dbf","Type":"ContainerDied","Data":"fd8145ab7dcf6d7a1d56c5b97a382302a90ddece38539fefd1b7a4258e63a35a"} Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.653372 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77f7bfffdc-cnbqk" event={"ID":"5352a37f-00de-4087-8337-6536a0c03fd5","Type":"ContainerStarted","Data":"9d961ffdbca1348bf88e054d0fe9890b289b17b4edb6a1b107a5f4aadb948695"} Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.658624 4725 generic.go:334] "Generic (PLEG): container finished" podID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerID="50f3d76cd46ff7389f603110bce47913ecba235795a58bff55804b656e6c4e01" exitCode=0 Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.658658 4725 generic.go:334] "Generic (PLEG): container finished" podID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerID="3b7f664409d34c4e26d27f826855e5d9ee8a05571c9af802547b8c790b55c94d" exitCode=143 Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.658660 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2bbca85-b50a-46b4-857e-6e3a2217b5da","Type":"ContainerDied","Data":"50f3d76cd46ff7389f603110bce47913ecba235795a58bff55804b656e6c4e01"} Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.658698 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2bbca85-b50a-46b4-857e-6e3a2217b5da","Type":"ContainerDied","Data":"3b7f664409d34c4e26d27f826855e5d9ee8a05571c9af802547b8c790b55c94d"} Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.661711 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-log" containerID="cri-o://0cfc71d4d21cd4c9402885507cfe656bf22e7953c00eb938c1887acdcf68cd81" gracePeriod=30 Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.661850 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a","Type":"ContainerStarted","Data":"66223e6d416177c7dccf48187f0bc9af4871007e26999728d7fe884c9a0ca4d8"} Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.661954 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-httpd" containerID="cri-o://66223e6d416177c7dccf48187f0bc9af4871007e26999728d7fe884c9a0ca4d8" gracePeriod=30 Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.724322 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.724304985 podStartE2EDuration="6.724304985s" podCreationTimestamp="2025-12-02 13:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:12.699212592 +0000 UTC m=+1063.655854297" watchObservedRunningTime="2025-12-02 13:22:12.724304985 +0000 UTC m=+1063.680946680" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.771376 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.949811 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z656k\" (UniqueName: \"kubernetes.io/projected/e2bbca85-b50a-46b4-857e-6e3a2217b5da-kube-api-access-z656k\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.949907 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.949976 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-scripts\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.950406 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-logs\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.950450 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-httpd-run\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.950530 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-config-data\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.950583 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-combined-ca-bundle\") pod \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\" (UID: \"e2bbca85-b50a-46b4-857e-6e3a2217b5da\") " Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.951090 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.951160 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-logs" (OuterVolumeSpecName: "logs") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.951435 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.951497 4725 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2bbca85-b50a-46b4-857e-6e3a2217b5da-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.957529 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-scripts" (OuterVolumeSpecName: "scripts") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.957728 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:22:12 crc kubenswrapper[4725]: I1202 13:22:12.957803 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2bbca85-b50a-46b4-857e-6e3a2217b5da-kube-api-access-z656k" (OuterVolumeSpecName: "kube-api-access-z656k") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "kube-api-access-z656k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.001548 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.033190 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-config-data" (OuterVolumeSpecName: "config-data") pod "e2bbca85-b50a-46b4-857e-6e3a2217b5da" (UID: "e2bbca85-b50a-46b4-857e-6e3a2217b5da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.053574 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z656k\" (UniqueName: \"kubernetes.io/projected/e2bbca85-b50a-46b4-857e-6e3a2217b5da-kube-api-access-z656k\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.053635 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.053650 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.053663 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.053676 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2bbca85-b50a-46b4-857e-6e3a2217b5da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.078361 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.158765 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.678276 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e2bbca85-b50a-46b4-857e-6e3a2217b5da","Type":"ContainerDied","Data":"953369eb2daef17e909e219a0feae590bc4a1db6a6560c33bb98608b1b2468d2"} Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.678329 4725 scope.go:117] "RemoveContainer" containerID="50f3d76cd46ff7389f603110bce47913ecba235795a58bff55804b656e6c4e01" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.678441 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.689184 4725 generic.go:334] "Generic (PLEG): container finished" podID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerID="66223e6d416177c7dccf48187f0bc9af4871007e26999728d7fe884c9a0ca4d8" exitCode=0 Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.689218 4725 generic.go:334] "Generic (PLEG): container finished" podID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerID="0cfc71d4d21cd4c9402885507cfe656bf22e7953c00eb938c1887acdcf68cd81" exitCode=143 Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.689243 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a","Type":"ContainerDied","Data":"66223e6d416177c7dccf48187f0bc9af4871007e26999728d7fe884c9a0ca4d8"} Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.689296 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a","Type":"ContainerDied","Data":"0cfc71d4d21cd4c9402885507cfe656bf22e7953c00eb938c1887acdcf68cd81"} Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.701096 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.708242 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.736438 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:13 crc kubenswrapper[4725]: E1202 13:22:13.736981 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-log" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.737006 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-log" Dec 02 13:22:13 crc kubenswrapper[4725]: E1202 13:22:13.737019 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-httpd" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.737029 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-httpd" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.737253 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-log" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.737279 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" containerName="glance-httpd" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.738924 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.741805 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.748001 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877409 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-logs\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877464 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877515 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877563 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877624 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877640 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.877668 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbbz2\" (UniqueName: \"kubernetes.io/projected/ca499ceb-5adb-4840-86e4-6fb053291cc3-kube-api-access-kbbz2\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.980517 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.980637 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.980736 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbbz2\" (UniqueName: \"kubernetes.io/projected/ca499ceb-5adb-4840-86e4-6fb053291cc3-kube-api-access-kbbz2\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.980838 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-logs\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.980890 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.980969 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.981089 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.981767 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.981825 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-logs\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.982211 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 02 13:22:13 crc kubenswrapper[4725]: I1202 13:22:13.990292 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-scripts\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:14 crc kubenswrapper[4725]: I1202 13:22:14.005764 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:14 crc kubenswrapper[4725]: I1202 13:22:14.010130 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbbz2\" (UniqueName: \"kubernetes.io/projected/ca499ceb-5adb-4840-86e4-6fb053291cc3-kube-api-access-kbbz2\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:14 crc kubenswrapper[4725]: I1202 13:22:14.026899 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-config-data\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:14 crc kubenswrapper[4725]: I1202 13:22:14.099177 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:14 crc kubenswrapper[4725]: I1202 13:22:14.375413 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:15 crc kubenswrapper[4725]: I1202 13:22:15.056107 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:15 crc kubenswrapper[4725]: I1202 13:22:15.279354 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2bbca85-b50a-46b4-857e-6e3a2217b5da" path="/var/lib/kubelet/pods/e2bbca85-b50a-46b4-857e-6e3a2217b5da/volumes" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.243796 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74f4bf4679-pfzbz"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.276082 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f44bd87bd-2w6jw"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.278583 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.281737 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.305329 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f44bd87bd-2w6jw"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332455 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-tls-certs\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332547 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-config-data\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332598 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-combined-ca-bundle\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332676 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kftlz\" (UniqueName: \"kubernetes.io/projected/cd102f85-1b92-4215-95e9-33baf5949dec-kube-api-access-kftlz\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332706 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd102f85-1b92-4215-95e9-33baf5949dec-logs\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332739 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-secret-key\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.332766 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-scripts\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.372731 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77f7bfffdc-cnbqk"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.403746 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f9f547644-22mp9"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.405807 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.422247 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f9f547644-22mp9"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.434980 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-combined-ca-bundle\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.435067 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kftlz\" (UniqueName: \"kubernetes.io/projected/cd102f85-1b92-4215-95e9-33baf5949dec-kube-api-access-kftlz\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.435094 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd102f85-1b92-4215-95e9-33baf5949dec-logs\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.435121 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-secret-key\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.435142 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-scripts\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.435170 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-tls-certs\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.435209 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-config-data\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.436170 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-scripts\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.436454 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-config-data\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.436585 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd102f85-1b92-4215-95e9-33baf5949dec-logs\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.440851 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-combined-ca-bundle\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.457890 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-secret-key\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.466129 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-tls-certs\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.466423 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kftlz\" (UniqueName: \"kubernetes.io/projected/cd102f85-1b92-4215-95e9-33baf5949dec-kube-api-access-kftlz\") pod \"horizon-7f44bd87bd-2w6jw\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.536952 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-horizon-tls-certs\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.539115 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-combined-ca-bundle\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.539402 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-logs\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.540681 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-horizon-secret-key\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.540746 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j47mw\" (UniqueName: \"kubernetes.io/projected/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-kube-api-access-j47mw\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.540769 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-scripts\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.540829 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-config-data\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.617561 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.642720 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-logs\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.643053 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-horizon-secret-key\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.643091 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j47mw\" (UniqueName: \"kubernetes.io/projected/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-kube-api-access-j47mw\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.643116 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-scripts\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.643508 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-config-data\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.643706 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-logs\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.644074 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-scripts\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.644610 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-horizon-tls-certs\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.644840 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-combined-ca-bundle\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.644871 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-config-data\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.647530 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-horizon-secret-key\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.648342 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-combined-ca-bundle\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.651207 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-horizon-tls-certs\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.659553 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j47mw\" (UniqueName: \"kubernetes.io/projected/a4cfe8aa-4818-4626-8e86-6607e7bb38d2-kube-api-access-j47mw\") pod \"horizon-f9f547644-22mp9\" (UID: \"a4cfe8aa-4818-4626-8e86-6607e7bb38d2\") " pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.727782 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.864622 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.888582 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.962973 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-credential-keys\") pod \"649811b5-8aff-4f44-8799-850000589dbf\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.963697 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6rw5\" (UniqueName: \"kubernetes.io/projected/649811b5-8aff-4f44-8799-850000589dbf-kube-api-access-c6rw5\") pod \"649811b5-8aff-4f44-8799-850000589dbf\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.963788 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-config-data\") pod \"649811b5-8aff-4f44-8799-850000589dbf\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.963862 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-fernet-keys\") pod \"649811b5-8aff-4f44-8799-850000589dbf\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.963924 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-scripts\") pod \"649811b5-8aff-4f44-8799-850000589dbf\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.963972 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-combined-ca-bundle\") pod \"649811b5-8aff-4f44-8799-850000589dbf\" (UID: \"649811b5-8aff-4f44-8799-850000589dbf\") " Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.969750 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/649811b5-8aff-4f44-8799-850000589dbf-kube-api-access-c6rw5" (OuterVolumeSpecName: "kube-api-access-c6rw5") pod "649811b5-8aff-4f44-8799-850000589dbf" (UID: "649811b5-8aff-4f44-8799-850000589dbf"). InnerVolumeSpecName "kube-api-access-c6rw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.979414 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-scripts" (OuterVolumeSpecName: "scripts") pod "649811b5-8aff-4f44-8799-850000589dbf" (UID: "649811b5-8aff-4f44-8799-850000589dbf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.979437 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "649811b5-8aff-4f44-8799-850000589dbf" (UID: "649811b5-8aff-4f44-8799-850000589dbf"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.982710 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmdk4"] Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.983595 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "649811b5-8aff-4f44-8799-850000589dbf" (UID: "649811b5-8aff-4f44-8799-850000589dbf"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:16 crc kubenswrapper[4725]: I1202 13:22:16.984426 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qmdk4" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" containerID="cri-o://a47074e6329a6a09eb6dcfff9b5be160af9ab43451062e9db24d19f7dd5aa947" gracePeriod=10 Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.029484 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "649811b5-8aff-4f44-8799-850000589dbf" (UID: "649811b5-8aff-4f44-8799-850000589dbf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.030863 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-config-data" (OuterVolumeSpecName: "config-data") pod "649811b5-8aff-4f44-8799-850000589dbf" (UID: "649811b5-8aff-4f44-8799-850000589dbf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.066771 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.066805 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.066816 4725 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.066829 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6rw5\" (UniqueName: \"kubernetes.io/projected/649811b5-8aff-4f44-8799-850000589dbf-kube-api-access-c6rw5\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.066837 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.066845 4725 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/649811b5-8aff-4f44-8799-850000589dbf-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.744020 4725 generic.go:334] "Generic (PLEG): container finished" podID="33e0067a-bebb-4abf-af23-42be9e914232" containerID="a47074e6329a6a09eb6dcfff9b5be160af9ab43451062e9db24d19f7dd5aa947" exitCode=0 Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.744105 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmdk4" event={"ID":"33e0067a-bebb-4abf-af23-42be9e914232","Type":"ContainerDied","Data":"a47074e6329a6a09eb6dcfff9b5be160af9ab43451062e9db24d19f7dd5aa947"} Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.746946 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b6q6j" event={"ID":"649811b5-8aff-4f44-8799-850000589dbf","Type":"ContainerDied","Data":"ced5c87f870a9ae316ec06cae78dd0c8bbe5437cd49d50a8008f37245edcb44c"} Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.746978 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ced5c87f870a9ae316ec06cae78dd0c8bbe5437cd49d50a8008f37245edcb44c" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.747020 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b6q6j" Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.942348 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b6q6j"] Dec 02 13:22:17 crc kubenswrapper[4725]: I1202 13:22:17.957640 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b6q6j"] Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.045847 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kzkrh"] Dec 02 13:22:18 crc kubenswrapper[4725]: E1202 13:22:18.046434 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649811b5-8aff-4f44-8799-850000589dbf" containerName="keystone-bootstrap" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.046452 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="649811b5-8aff-4f44-8799-850000589dbf" containerName="keystone-bootstrap" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.046682 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="649811b5-8aff-4f44-8799-850000589dbf" containerName="keystone-bootstrap" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.047395 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.049711 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.056508 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kzkrh"] Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.056752 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.057667 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-twszr" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.057869 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.058085 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.188392 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-combined-ca-bundle\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.188453 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-scripts\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.188548 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-config-data\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.188650 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-fernet-keys\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.188711 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hsnm\" (UniqueName: \"kubernetes.io/projected/3128bc10-71dd-46f4-8e71-787260ede3d0-kube-api-access-9hsnm\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.188855 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-credential-keys\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.289952 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-config-data\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.290015 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-fernet-keys\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.290042 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hsnm\" (UniqueName: \"kubernetes.io/projected/3128bc10-71dd-46f4-8e71-787260ede3d0-kube-api-access-9hsnm\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.290131 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-credential-keys\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.290181 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-combined-ca-bundle\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.290199 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-scripts\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.295093 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-scripts\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.295424 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-config-data\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.295485 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-fernet-keys\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.298932 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-combined-ca-bundle\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.306085 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-credential-keys\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.307895 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hsnm\" (UniqueName: \"kubernetes.io/projected/3128bc10-71dd-46f4-8e71-787260ede3d0-kube-api-access-9hsnm\") pod \"keystone-bootstrap-kzkrh\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:18 crc kubenswrapper[4725]: I1202 13:22:18.376741 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:19 crc kubenswrapper[4725]: I1202 13:22:19.276991 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="649811b5-8aff-4f44-8799-850000589dbf" path="/var/lib/kubelet/pods/649811b5-8aff-4f44-8799-850000589dbf/volumes" Dec 02 13:22:20 crc kubenswrapper[4725]: I1202 13:22:20.569409 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qmdk4" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.267452 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.363975 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.364083 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-combined-ca-bundle\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.364156 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-config-data\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.364219 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdqhk\" (UniqueName: \"kubernetes.io/projected/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-kube-api-access-sdqhk\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.364333 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-httpd-run\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.365397 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.365490 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-scripts\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.365531 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-logs\") pod \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\" (UID: \"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a\") " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.365936 4725 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.366633 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-logs" (OuterVolumeSpecName: "logs") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.371509 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.374123 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-kube-api-access-sdqhk" (OuterVolumeSpecName: "kube-api-access-sdqhk") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "kube-api-access-sdqhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.374251 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-scripts" (OuterVolumeSpecName: "scripts") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.401632 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.423380 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-config-data" (OuterVolumeSpecName: "config-data") pod "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" (UID: "99b94af9-f4c3-4acd-bb4c-50d54ecaae4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.467785 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.467822 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdqhk\" (UniqueName: \"kubernetes.io/projected/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-kube-api-access-sdqhk\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.467833 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.467843 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.467873 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.467883 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.487624 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.570214 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.793269 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99b94af9-f4c3-4acd-bb4c-50d54ecaae4a","Type":"ContainerDied","Data":"77c84d5f2003f710e37cbe8807629d297489eb10373ddefc1a7e63719b644d7c"} Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.793350 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.835283 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.852134 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.864343 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:22 crc kubenswrapper[4725]: E1202 13:22:22.865262 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-log" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.865281 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-log" Dec 02 13:22:22 crc kubenswrapper[4725]: E1202 13:22:22.865302 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-httpd" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.865309 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-httpd" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.865501 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-httpd" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.865527 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" containerName="glance-log" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.866886 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.868793 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.870222 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.874985 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.980836 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.980957 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.981115 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.981192 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.981321 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccl4r\" (UniqueName: \"kubernetes.io/projected/153f44c9-3565-45bc-a162-8b7646df9cb2-kube-api-access-ccl4r\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.981386 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.981513 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-logs\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:22 crc kubenswrapper[4725]: I1202 13:22:22.981583 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.082862 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083253 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083362 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccl4r\" (UniqueName: \"kubernetes.io/projected/153f44c9-3565-45bc-a162-8b7646df9cb2-kube-api-access-ccl4r\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083381 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083402 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-logs\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083421 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083445 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083502 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083790 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.083933 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.084183 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-logs\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.088953 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.089174 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.090813 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.090837 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.102102 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccl4r\" (UniqueName: \"kubernetes.io/projected/153f44c9-3565-45bc-a162-8b7646df9cb2-kube-api-access-ccl4r\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.123231 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.196157 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:23 crc kubenswrapper[4725]: I1202 13:22:23.278503 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b94af9-f4c3-4acd-bb4c-50d54ecaae4a" path="/var/lib/kubelet/pods/99b94af9-f4c3-4acd-bb4c-50d54ecaae4a/volumes" Dec 02 13:22:25 crc kubenswrapper[4725]: I1202 13:22:25.569404 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qmdk4" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 02 13:22:35 crc kubenswrapper[4725]: I1202 13:22:35.477689 4725 scope.go:117] "RemoveContainer" containerID="3b7f664409d34c4e26d27f826855e5d9ee8a05571c9af802547b8c790b55c94d" Dec 02 13:22:35 crc kubenswrapper[4725]: I1202 13:22:35.569396 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qmdk4" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 02 13:22:35 crc kubenswrapper[4725]: I1202 13:22:35.569999 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:22:36 crc kubenswrapper[4725]: E1202 13:22:36.000949 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 02 13:22:36 crc kubenswrapper[4725]: E1202 13:22:36.001357 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z9nbc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-9blbh_openstack(52677bda-0bda-4d69-af01-f6ce198582f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:22:36 crc kubenswrapper[4725]: E1202 13:22:36.002605 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-9blbh" podUID="52677bda-0bda-4d69-af01-f6ce198582f6" Dec 02 13:22:36 crc kubenswrapper[4725]: E1202 13:22:36.338066 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 02 13:22:36 crc kubenswrapper[4725]: E1202 13:22:36.338255 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n655h69hbdh57h68h59bh548h67bh5fch8dhf9h5dfhf7h5fbh565h586h645h595h667h679h65dh4h59ch557hfch67bh66fh597h67dh56fh64ch588q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5r6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(335fa14a-e2ec-4f9e-a56d-ab2a9be708f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.410350 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.488836 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/33e0067a-bebb-4abf-af23-42be9e914232-kube-api-access-qx9db\") pod \"33e0067a-bebb-4abf-af23-42be9e914232\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.489051 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-dns-svc\") pod \"33e0067a-bebb-4abf-af23-42be9e914232\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.489108 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-nb\") pod \"33e0067a-bebb-4abf-af23-42be9e914232\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.489208 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-config\") pod \"33e0067a-bebb-4abf-af23-42be9e914232\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.489254 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-sb\") pod \"33e0067a-bebb-4abf-af23-42be9e914232\" (UID: \"33e0067a-bebb-4abf-af23-42be9e914232\") " Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.494876 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e0067a-bebb-4abf-af23-42be9e914232-kube-api-access-qx9db" (OuterVolumeSpecName: "kube-api-access-qx9db") pod "33e0067a-bebb-4abf-af23-42be9e914232" (UID: "33e0067a-bebb-4abf-af23-42be9e914232"). InnerVolumeSpecName "kube-api-access-qx9db". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.532734 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "33e0067a-bebb-4abf-af23-42be9e914232" (UID: "33e0067a-bebb-4abf-af23-42be9e914232"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.532880 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "33e0067a-bebb-4abf-af23-42be9e914232" (UID: "33e0067a-bebb-4abf-af23-42be9e914232"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.546316 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "33e0067a-bebb-4abf-af23-42be9e914232" (UID: "33e0067a-bebb-4abf-af23-42be9e914232"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.554273 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-config" (OuterVolumeSpecName: "config") pod "33e0067a-bebb-4abf-af23-42be9e914232" (UID: "33e0067a-bebb-4abf-af23-42be9e914232"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.591415 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.591453 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.591488 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.591498 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/33e0067a-bebb-4abf-af23-42be9e914232-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.591511 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx9db\" (UniqueName: \"kubernetes.io/projected/33e0067a-bebb-4abf-af23-42be9e914232-kube-api-access-qx9db\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.960859 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qmdk4" Dec 02 13:22:36 crc kubenswrapper[4725]: I1202 13:22:36.960915 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qmdk4" event={"ID":"33e0067a-bebb-4abf-af23-42be9e914232","Type":"ContainerDied","Data":"2a779da4a98fdbb55f59071416abbfc18764637bf097e2efb7d981a537664d4f"} Dec 02 13:22:36 crc kubenswrapper[4725]: E1202 13:22:36.962792 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-9blbh" podUID="52677bda-0bda-4d69-af01-f6ce198582f6" Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.012112 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmdk4"] Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.023322 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qmdk4"] Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.278251 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33e0067a-bebb-4abf-af23-42be9e914232" path="/var/lib/kubelet/pods/33e0067a-bebb-4abf-af23-42be9e914232/volumes" Dec 02 13:22:37 crc kubenswrapper[4725]: E1202 13:22:37.525645 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 02 13:22:37 crc kubenswrapper[4725]: E1202 13:22:37.526057 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-96b4p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-pdwcp_openstack(11048035-cfbd-4e03-8427-83e0d2a91a63): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:22:37 crc kubenswrapper[4725]: E1202 13:22:37.527237 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-pdwcp" podUID="11048035-cfbd-4e03-8427-83e0d2a91a63" Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.552277 4725 scope.go:117] "RemoveContainer" containerID="66223e6d416177c7dccf48187f0bc9af4871007e26999728d7fe884c9a0ca4d8" Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.727217 4725 scope.go:117] "RemoveContainer" containerID="0cfc71d4d21cd4c9402885507cfe656bf22e7953c00eb938c1887acdcf68cd81" Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.805821 4725 scope.go:117] "RemoveContainer" containerID="a47074e6329a6a09eb6dcfff9b5be160af9ab43451062e9db24d19f7dd5aa947" Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.846251 4725 scope.go:117] "RemoveContainer" containerID="d5aabec06ed4d824f780c8ec733dceecc410de95e0198229412d107b7bb213c4" Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.981755 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77f7bfffdc-cnbqk" event={"ID":"5352a37f-00de-4087-8337-6536a0c03fd5","Type":"ContainerStarted","Data":"1e73159d6ea8897fc75b41f9b67e58dc11bc565d76fa18bd6fad3abfa225b947"} Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.985375 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5489f8c44c-g22fg" event={"ID":"de526deb-a82c-469a-9387-308555f0d667","Type":"ContainerStarted","Data":"73f8e7a7c9725ec6b7f049eec81a1ae192748e5d7b3c158a640ccbf9c18d217f"} Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.990564 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f4bf4679-pfzbz" event={"ID":"df2ed35d-e2f2-4c5f-b2db-12d95291d167","Type":"ContainerStarted","Data":"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b"} Dec 02 13:22:37 crc kubenswrapper[4725]: I1202 13:22:37.993611 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-psww9" event={"ID":"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e","Type":"ContainerStarted","Data":"1b967543fe877f2d2628c57a87a3c6aeccf0ccb6407d44af3ec41221292a1f16"} Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.014170 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-psww9" podStartSLOduration=3.5232513279999997 podStartE2EDuration="32.014145296s" podCreationTimestamp="2025-12-02 13:22:06 +0000 UTC" firstStartedPulling="2025-12-02 13:22:08.9916562 +0000 UTC m=+1059.948297895" lastFinishedPulling="2025-12-02 13:22:37.482550168 +0000 UTC m=+1088.439191863" observedRunningTime="2025-12-02 13:22:38.00944312 +0000 UTC m=+1088.966084815" watchObservedRunningTime="2025-12-02 13:22:38.014145296 +0000 UTC m=+1088.970786991" Dec 02 13:22:38 crc kubenswrapper[4725]: E1202 13:22:38.018112 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-pdwcp" podUID="11048035-cfbd-4e03-8427-83e0d2a91a63" Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.071259 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f44bd87bd-2w6jw"] Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.186960 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kzkrh"] Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.221527 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.236199 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.265497 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f9f547644-22mp9"] Dec 02 13:22:38 crc kubenswrapper[4725]: I1202 13:22:38.333632 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:22:38 crc kubenswrapper[4725]: W1202 13:22:38.890352 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod153f44c9_3565_45bc_a162_8b7646df9cb2.slice/crio-3a0bedf12eee23f5f1c996d4bfddf3fcad9ee79eb97cd73a24d1fe46a3652a9d WatchSource:0}: Error finding container 3a0bedf12eee23f5f1c996d4bfddf3fcad9ee79eb97cd73a24d1fe46a3652a9d: Status 404 returned error can't find the container with id 3a0bedf12eee23f5f1c996d4bfddf3fcad9ee79eb97cd73a24d1fe46a3652a9d Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.039064 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ca499ceb-5adb-4840-86e4-6fb053291cc3","Type":"ContainerStarted","Data":"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.039645 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ca499ceb-5adb-4840-86e4-6fb053291cc3","Type":"ContainerStarted","Data":"de930a4304dda4aabeb8cdcd3ed9caf5a02c81142bf61dc43f8b30c7be59ec28"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.052220 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77f7bfffdc-cnbqk" event={"ID":"5352a37f-00de-4087-8337-6536a0c03fd5","Type":"ContainerStarted","Data":"c2a0785f7ce3393a1b5e7ddbe3ba5ef2b58150b80527a0ce28d3f98863104a5d"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.052453 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77f7bfffdc-cnbqk" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon-log" containerID="cri-o://1e73159d6ea8897fc75b41f9b67e58dc11bc565d76fa18bd6fad3abfa225b947" gracePeriod=30 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.053124 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77f7bfffdc-cnbqk" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon" containerID="cri-o://c2a0785f7ce3393a1b5e7ddbe3ba5ef2b58150b80527a0ce28d3f98863104a5d" gracePeriod=30 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.064943 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f9f547644-22mp9" event={"ID":"a4cfe8aa-4818-4626-8e86-6607e7bb38d2","Type":"ContainerStarted","Data":"a93c9f8877a9550006e43e2fe9d917bd83e5155e4c47efe28faffec6c6ca0394"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.076642 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77f7bfffdc-cnbqk" podStartSLOduration=2.901070923 podStartE2EDuration="28.076579795s" podCreationTimestamp="2025-12-02 13:22:11 +0000 UTC" firstStartedPulling="2025-12-02 13:22:12.402719941 +0000 UTC m=+1063.359361636" lastFinishedPulling="2025-12-02 13:22:37.578228803 +0000 UTC m=+1088.534870508" observedRunningTime="2025-12-02 13:22:39.072987786 +0000 UTC m=+1090.029629481" watchObservedRunningTime="2025-12-02 13:22:39.076579795 +0000 UTC m=+1090.033221490" Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.078790 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5489f8c44c-g22fg" event={"ID":"de526deb-a82c-469a-9387-308555f0d667","Type":"ContainerStarted","Data":"f8e3cd60233f0e0901e1480751dc5a6c7379feeb0fe885446b85d4ff1b41ff30"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.079262 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5489f8c44c-g22fg" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon" containerID="cri-o://f8e3cd60233f0e0901e1480751dc5a6c7379feeb0fe885446b85d4ff1b41ff30" gracePeriod=30 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.079347 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5489f8c44c-g22fg" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon-log" containerID="cri-o://73f8e7a7c9725ec6b7f049eec81a1ae192748e5d7b3c158a640ccbf9c18d217f" gracePeriod=30 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.086966 4725 generic.go:334] "Generic (PLEG): container finished" podID="b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" containerID="39ea52ea6285c01b3b519293f937ef8f9d19b4e72bc46fa935f7e64f87e1b08d" exitCode=0 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.087080 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lx9dr" event={"ID":"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76","Type":"ContainerDied","Data":"39ea52ea6285c01b3b519293f937ef8f9d19b4e72bc46fa935f7e64f87e1b08d"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.089134 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f4bf4679-pfzbz" event={"ID":"df2ed35d-e2f2-4c5f-b2db-12d95291d167","Type":"ContainerStarted","Data":"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.089272 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74f4bf4679-pfzbz" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon-log" containerID="cri-o://0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b" gracePeriod=30 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.089451 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74f4bf4679-pfzbz" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon" containerID="cri-o://b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373" gracePeriod=30 Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.092036 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzkrh" event={"ID":"3128bc10-71dd-46f4-8e71-787260ede3d0","Type":"ContainerStarted","Data":"c2db4ef7011627e7305823e5fc06fa145a1cbf465669cb3cd6c90792999b81e8"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.098929 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44bd87bd-2w6jw" event={"ID":"cd102f85-1b92-4215-95e9-33baf5949dec","Type":"ContainerStarted","Data":"149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.098977 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44bd87bd-2w6jw" event={"ID":"cd102f85-1b92-4215-95e9-33baf5949dec","Type":"ContainerStarted","Data":"c9f9992cd5ffd2a3837f9f5d8a78714847a99fd4af568dd4010844d47566a212"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.108399 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5489f8c44c-g22fg" podStartSLOduration=4.77737923 podStartE2EDuration="34.108369924s" podCreationTimestamp="2025-12-02 13:22:05 +0000 UTC" firstStartedPulling="2025-12-02 13:22:08.254646973 +0000 UTC m=+1059.211288668" lastFinishedPulling="2025-12-02 13:22:37.585637667 +0000 UTC m=+1088.542279362" observedRunningTime="2025-12-02 13:22:39.099491354 +0000 UTC m=+1090.056133049" watchObservedRunningTime="2025-12-02 13:22:39.108369924 +0000 UTC m=+1090.065011619" Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.133857 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"153f44c9-3565-45bc-a162-8b7646df9cb2","Type":"ContainerStarted","Data":"3a0bedf12eee23f5f1c996d4bfddf3fcad9ee79eb97cd73a24d1fe46a3652a9d"} Dec 02 13:22:39 crc kubenswrapper[4725]: I1202 13:22:39.146612 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74f4bf4679-pfzbz" podStartSLOduration=5.842943812 podStartE2EDuration="33.146591123s" podCreationTimestamp="2025-12-02 13:22:06 +0000 UTC" firstStartedPulling="2025-12-02 13:22:09.01501634 +0000 UTC m=+1059.971658035" lastFinishedPulling="2025-12-02 13:22:36.318663651 +0000 UTC m=+1087.275305346" observedRunningTime="2025-12-02 13:22:39.141327453 +0000 UTC m=+1090.097969148" watchObservedRunningTime="2025-12-02 13:22:39.146591123 +0000 UTC m=+1090.103232818" Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.176573 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzkrh" event={"ID":"3128bc10-71dd-46f4-8e71-787260ede3d0","Type":"ContainerStarted","Data":"defad413a0b029b7aa4a113098e7b7a85fd86214d0d36a873dc97ae4d29e840d"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.182114 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44bd87bd-2w6jw" event={"ID":"cd102f85-1b92-4215-95e9-33baf5949dec","Type":"ContainerStarted","Data":"0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.202637 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kzkrh" podStartSLOduration=22.202618983 podStartE2EDuration="22.202618983s" podCreationTimestamp="2025-12-02 13:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:40.200807878 +0000 UTC m=+1091.157449573" watchObservedRunningTime="2025-12-02 13:22:40.202618983 +0000 UTC m=+1091.159260678" Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.208871 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"153f44c9-3565-45bc-a162-8b7646df9cb2","Type":"ContainerStarted","Data":"6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.217760 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ca499ceb-5adb-4840-86e4-6fb053291cc3","Type":"ContainerStarted","Data":"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.217944 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-log" containerID="cri-o://25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0" gracePeriod=30 Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.218379 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-httpd" containerID="cri-o://7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b" gracePeriod=30 Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.231404 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f9f547644-22mp9" event={"ID":"a4cfe8aa-4818-4626-8e86-6607e7bb38d2","Type":"ContainerStarted","Data":"bcced6c74d5b61e013fcf1b55d8b7a3a20e82ac4ba9ea76d0245a8dc9db61be2"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.231448 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f9f547644-22mp9" event={"ID":"a4cfe8aa-4818-4626-8e86-6607e7bb38d2","Type":"ContainerStarted","Data":"bb1ee01cf342f711979f6edc1dd5f7d501569f2bfa90a50c6057c8aee6af497c"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.239585 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7f44bd87bd-2w6jw" podStartSLOduration=24.239481748 podStartE2EDuration="24.239481748s" podCreationTimestamp="2025-12-02 13:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:40.231076089 +0000 UTC m=+1091.187717784" watchObservedRunningTime="2025-12-02 13:22:40.239481748 +0000 UTC m=+1091.196123443" Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.241006 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerStarted","Data":"2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc"} Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.267035 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.266997471 podStartE2EDuration="27.266997471s" podCreationTimestamp="2025-12-02 13:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:40.251967878 +0000 UTC m=+1091.208609563" watchObservedRunningTime="2025-12-02 13:22:40.266997471 +0000 UTC m=+1091.223639166" Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.285402 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f9f547644-22mp9" podStartSLOduration=24.285379138 podStartE2EDuration="24.285379138s" podCreationTimestamp="2025-12-02 13:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:40.278966768 +0000 UTC m=+1091.235608473" watchObservedRunningTime="2025-12-02 13:22:40.285379138 +0000 UTC m=+1091.242020833" Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.573135 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qmdk4" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 02 13:22:40 crc kubenswrapper[4725]: I1202 13:22:40.890208 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.045526 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-combined-ca-bundle\") pod \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.045897 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mppxv\" (UniqueName: \"kubernetes.io/projected/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-kube-api-access-mppxv\") pod \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.045961 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-config\") pod \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\" (UID: \"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.054953 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-kube-api-access-mppxv" (OuterVolumeSpecName: "kube-api-access-mppxv") pod "b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" (UID: "b3ed1edd-ee3b-48a5-b4a6-539e780a2e76"). InnerVolumeSpecName "kube-api-access-mppxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.082108 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" (UID: "b3ed1edd-ee3b-48a5-b4a6-539e780a2e76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.111449 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-config" (OuterVolumeSpecName: "config") pod "b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" (UID: "b3ed1edd-ee3b-48a5-b4a6-539e780a2e76"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.147907 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mppxv\" (UniqueName: \"kubernetes.io/projected/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-kube-api-access-mppxv\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.147946 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.147960 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.166006 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.256497 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lx9dr" event={"ID":"b3ed1edd-ee3b-48a5-b4a6-539e780a2e76","Type":"ContainerDied","Data":"34e6b07e25c1fb2c0b4892de6b2ce5f2ee4937293303949b761bd9752cd24c56"} Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.256540 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34e6b07e25c1fb2c0b4892de6b2ce5f2ee4937293303949b761bd9752cd24c56" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.256541 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lx9dr" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.264152 4725 generic.go:334] "Generic (PLEG): container finished" podID="b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" containerID="1b967543fe877f2d2628c57a87a3c6aeccf0ccb6407d44af3ec41221292a1f16" exitCode=0 Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.264566 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-psww9" event={"ID":"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e","Type":"ContainerDied","Data":"1b967543fe877f2d2628c57a87a3c6aeccf0ccb6407d44af3ec41221292a1f16"} Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.282452 4725 generic.go:334] "Generic (PLEG): container finished" podID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerID="7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b" exitCode=143 Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.282662 4725 generic.go:334] "Generic (PLEG): container finished" podID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerID="25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0" exitCode=143 Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.292323 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.392152 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"153f44c9-3565-45bc-a162-8b7646df9cb2","Type":"ContainerStarted","Data":"568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9"} Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.392209 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ca499ceb-5adb-4840-86e4-6fb053291cc3","Type":"ContainerDied","Data":"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b"} Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.392238 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ca499ceb-5adb-4840-86e4-6fb053291cc3","Type":"ContainerDied","Data":"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0"} Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.392255 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ca499ceb-5adb-4840-86e4-6fb053291cc3","Type":"ContainerDied","Data":"de930a4304dda4aabeb8cdcd3ed9caf5a02c81142bf61dc43f8b30c7be59ec28"} Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.392279 4725 scope.go:117] "RemoveContainer" containerID="7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.436356 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=19.436336773 podStartE2EDuration="19.436336773s" podCreationTimestamp="2025-12-02 13:22:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:41.409914767 +0000 UTC m=+1092.366556452" watchObservedRunningTime="2025-12-02 13:22:41.436336773 +0000 UTC m=+1092.392978468" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.463660 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.463741 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-httpd-run\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.463798 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-logs\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.463851 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbbz2\" (UniqueName: \"kubernetes.io/projected/ca499ceb-5adb-4840-86e4-6fb053291cc3-kube-api-access-kbbz2\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.463926 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-scripts\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.463950 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-combined-ca-bundle\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.464061 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-config-data\") pod \"ca499ceb-5adb-4840-86e4-6fb053291cc3\" (UID: \"ca499ceb-5adb-4840-86e4-6fb053291cc3\") " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.464588 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-logs" (OuterVolumeSpecName: "logs") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.464853 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.472827 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca499ceb-5adb-4840-86e4-6fb053291cc3-kube-api-access-kbbz2" (OuterVolumeSpecName: "kube-api-access-kbbz2") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "kube-api-access-kbbz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.485274 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vmglp"] Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.485733 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="init" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.485752 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="init" Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.485784 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-log" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.485791 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-log" Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.485803 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-httpd" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.485809 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-httpd" Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.485817 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" containerName="neutron-db-sync" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.485822 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" containerName="neutron-db-sync" Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.485833 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.485839 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.486000 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-log" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.486012 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e0067a-bebb-4abf-af23-42be9e914232" containerName="dnsmasq-dns" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.486026 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" containerName="neutron-db-sync" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.486035 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" containerName="glance-httpd" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.486929 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.486967 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-scripts" (OuterVolumeSpecName: "scripts") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.498824 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vmglp"] Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.508622 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.537303 4725 scope.go:117] "RemoveContainer" containerID="25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.555859 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.566111 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.566151 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.566184 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.566196 4725 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.566208 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca499ceb-5adb-4840-86e4-6fb053291cc3-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.566218 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbbz2\" (UniqueName: \"kubernetes.io/projected/ca499ceb-5adb-4840-86e4-6fb053291cc3-kube-api-access-kbbz2\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.604751 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64f9c49f9b-5bjhh"] Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.617484 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64f9c49f9b-5bjhh"] Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.617605 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.621322 4725 scope.go:117] "RemoveContainer" containerID="7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b" Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.621713 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b\": container with ID starting with 7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b not found: ID does not exist" containerID="7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.621752 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b"} err="failed to get container status \"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b\": rpc error: code = NotFound desc = could not find container \"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b\": container with ID starting with 7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b not found: ID does not exist" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.621774 4725 scope.go:117] "RemoveContainer" containerID="25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0" Dec 02 13:22:41 crc kubenswrapper[4725]: E1202 13:22:41.622393 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0\": container with ID starting with 25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0 not found: ID does not exist" containerID="25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.622428 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0"} err="failed to get container status \"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0\": rpc error: code = NotFound desc = could not find container \"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0\": container with ID starting with 25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0 not found: ID does not exist" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.622443 4725 scope.go:117] "RemoveContainer" containerID="7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.624721 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.626668 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b"} err="failed to get container status \"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b\": rpc error: code = NotFound desc = could not find container \"7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b\": container with ID starting with 7f7f3085878518fa221faea1e69205c0df1ec8866d5c8fd69801e02c35f8928b not found: ID does not exist" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.626694 4725 scope.go:117] "RemoveContainer" containerID="25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.626874 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.627174 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.627318 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.628800 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-wmgxq" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.635539 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0"} err="failed to get container status \"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0\": rpc error: code = NotFound desc = could not find container \"25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0\": container with ID starting with 25ff1ddf861a53d3d761da3bf704f5c59016d0c0e6f17a4373a092b03c65b0b0 not found: ID does not exist" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.635603 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-config-data" (OuterVolumeSpecName: "config-data") pod "ca499ceb-5adb-4840-86e4-6fb053291cc3" (UID: "ca499ceb-5adb-4840-86e4-6fb053291cc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.646950 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668507 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668572 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668685 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668776 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z5zz\" (UniqueName: \"kubernetes.io/projected/089c681d-b6d5-43ac-9fca-ed2a92d59441-kube-api-access-5z5zz\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668812 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-config\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668835 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668930 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca499ceb-5adb-4840-86e4-6fb053291cc3-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.668942 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770476 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770525 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z5zz\" (UniqueName: \"kubernetes.io/projected/089c681d-b6d5-43ac-9fca-ed2a92d59441-kube-api-access-5z5zz\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770551 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-config\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770571 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770598 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-ovndb-tls-certs\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770651 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-combined-ca-bundle\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770705 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8jvx\" (UniqueName: \"kubernetes.io/projected/66e1f474-e66c-4085-b7de-4f6ef80bfa84-kube-api-access-t8jvx\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770745 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770772 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770800 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-httpd-config\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.770841 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-config\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.772157 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.772739 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.772741 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.772942 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.773865 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-config\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.794498 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z5zz\" (UniqueName: \"kubernetes.io/projected/089c681d-b6d5-43ac-9fca-ed2a92d59441-kube-api-access-5z5zz\") pod \"dnsmasq-dns-55f844cf75-vmglp\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.876632 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-ovndb-tls-certs\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.876732 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-combined-ca-bundle\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.876799 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8jvx\" (UniqueName: \"kubernetes.io/projected/66e1f474-e66c-4085-b7de-4f6ef80bfa84-kube-api-access-t8jvx\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.876864 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-httpd-config\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.877359 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-config\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.880632 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-ovndb-tls-certs\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.881681 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-httpd-config\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.882105 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-combined-ca-bundle\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.882894 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-config\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.901256 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8jvx\" (UniqueName: \"kubernetes.io/projected/66e1f474-e66c-4085-b7de-4f6ef80bfa84-kube-api-access-t8jvx\") pod \"neutron-64f9c49f9b-5bjhh\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.939518 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.943703 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.950644 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.962836 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.982729 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.984158 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.989204 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 13:22:41 crc kubenswrapper[4725]: I1202 13:22:41.990988 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.001804 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.083877 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-config-data\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084393 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084446 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084490 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-scripts\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084579 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-logs\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084608 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084659 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bgsc\" (UniqueName: \"kubernetes.io/projected/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-kube-api-access-2bgsc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.084715 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.187597 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.187677 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-config-data\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.187727 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.187756 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.188636 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-scripts\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.188746 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-logs\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.188774 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.188807 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bgsc\" (UniqueName: \"kubernetes.io/projected/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-kube-api-access-2bgsc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.189907 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.193506 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.196364 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-logs\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.207386 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.207527 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.208991 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-scripts\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.210544 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-config-data\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.236335 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bgsc\" (UniqueName: \"kubernetes.io/projected/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-kube-api-access-2bgsc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.259667 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.309145 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.782201 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vmglp"] Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.806951 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64f9c49f9b-5bjhh"] Dec 02 13:22:42 crc kubenswrapper[4725]: W1202 13:22:42.848110 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66e1f474_e66c_4085_b7de_4f6ef80bfa84.slice/crio-7ea24c92f55e14f352638872ef5ce8eab1e1aea12abe60db0aa7624e9e502906 WatchSource:0}: Error finding container 7ea24c92f55e14f352638872ef5ce8eab1e1aea12abe60db0aa7624e9e502906: Status 404 returned error can't find the container with id 7ea24c92f55e14f352638872ef5ce8eab1e1aea12abe60db0aa7624e9e502906 Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.877818 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-psww9" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.916130 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-config-data\") pod \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.916217 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgxld\" (UniqueName: \"kubernetes.io/projected/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-kube-api-access-lgxld\") pod \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.916381 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-logs\") pod \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.916440 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-scripts\") pod \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.916489 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-combined-ca-bundle\") pod \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\" (UID: \"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e\") " Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.924024 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-logs" (OuterVolumeSpecName: "logs") pod "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" (UID: "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.940740 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-scripts" (OuterVolumeSpecName: "scripts") pod "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" (UID: "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.954733 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-kube-api-access-lgxld" (OuterVolumeSpecName: "kube-api-access-lgxld") pod "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" (UID: "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e"). InnerVolumeSpecName "kube-api-access-lgxld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:42 crc kubenswrapper[4725]: I1202 13:22:42.959753 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-config-data" (OuterVolumeSpecName: "config-data") pod "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" (UID: "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.017373 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" (UID: "b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.018740 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.018849 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgxld\" (UniqueName: \"kubernetes.io/projected/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-kube-api-access-lgxld\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.018937 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.019000 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.019053 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.200117 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.200653 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.391104 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca499ceb-5adb-4840-86e4-6fb053291cc3" path="/var/lib/kubelet/pods/ca499ceb-5adb-4840-86e4-6fb053291cc3/volumes" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.392251 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.445860 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-psww9" event={"ID":"b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e","Type":"ContainerDied","Data":"7eecbf2783c4d01dea1be3a0b5725341ce4a1ab1d5fd14e66878d24112b90e65"} Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.445903 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eecbf2783c4d01dea1be3a0b5725341ce4a1ab1d5fd14e66878d24112b90e65" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.445973 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-psww9" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.463793 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" event={"ID":"089c681d-b6d5-43ac-9fca-ed2a92d59441","Type":"ContainerStarted","Data":"04500d9780fc70441529ef6d545d41a54f52decbf79da38b5a08c0b19f8959aa"} Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.474146 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f9c49f9b-5bjhh" event={"ID":"66e1f474-e66c-4085-b7de-4f6ef80bfa84","Type":"ContainerStarted","Data":"7ea24c92f55e14f352638872ef5ce8eab1e1aea12abe60db0aa7624e9e502906"} Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.474191 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.482218 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.503344 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.564769 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-784f98d458-wsm8f"] Dec 02 13:22:43 crc kubenswrapper[4725]: E1202 13:22:43.574510 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" containerName="placement-db-sync" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.574648 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" containerName="placement-db-sync" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.587089 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" containerName="placement-db-sync" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.588565 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.593037 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.593243 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.593370 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.593415 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.593376 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-jqqg2" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.597070 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-784f98d458-wsm8f"] Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690126 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-public-tls-certs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690175 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-config-data\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690207 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-scripts\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690252 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66a69b36-610f-4f99-b57b-834d0a8a07a1-logs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690291 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-internal-tls-certs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690315 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtz4h\" (UniqueName: \"kubernetes.io/projected/66a69b36-610f-4f99-b57b-834d0a8a07a1-kube-api-access-wtz4h\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.690359 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-combined-ca-bundle\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.792738 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-internal-tls-certs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.792852 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtz4h\" (UniqueName: \"kubernetes.io/projected/66a69b36-610f-4f99-b57b-834d0a8a07a1-kube-api-access-wtz4h\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.792954 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-combined-ca-bundle\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.793044 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-public-tls-certs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.793078 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-config-data\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.793133 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-scripts\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.793216 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66a69b36-610f-4f99-b57b-834d0a8a07a1-logs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.793855 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66a69b36-610f-4f99-b57b-834d0a8a07a1-logs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.797303 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-public-tls-certs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.800067 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-internal-tls-certs\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.800839 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-scripts\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.808383 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-combined-ca-bundle\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.808448 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66a69b36-610f-4f99-b57b-834d0a8a07a1-config-data\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.818040 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtz4h\" (UniqueName: \"kubernetes.io/projected/66a69b36-610f-4f99-b57b-834d0a8a07a1-kube-api-access-wtz4h\") pod \"placement-784f98d458-wsm8f\" (UID: \"66a69b36-610f-4f99-b57b-834d0a8a07a1\") " pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:43 crc kubenswrapper[4725]: I1202 13:22:43.839439 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.502699 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f9c49f9b-5bjhh" event={"ID":"66e1f474-e66c-4085-b7de-4f6ef80bfa84","Type":"ContainerStarted","Data":"aa9195c8c0f9f6caee04b69871fa844b4d1778256decd817af290cf37baa99ed"} Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.503270 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f9c49f9b-5bjhh" event={"ID":"66e1f474-e66c-4085-b7de-4f6ef80bfa84","Type":"ContainerStarted","Data":"4ab58d5470d7602b49cc92f4e2bcbf04039ed1a49a679485db252829c8ab31cf"} Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.504593 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.509575 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29ba5625-386c-4e5f-8aa1-7ed349af1d7a","Type":"ContainerStarted","Data":"362bfba2783ee3da80abcdbf57b938ebf5cdd43fd4587798d8af88a39089b2ee"} Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.531006 4725 generic.go:334] "Generic (PLEG): container finished" podID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerID="5c8d5fff6a51feb41575531bc7f2c2030eaa343034944857c8509c81bdc2bdb2" exitCode=0 Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.531921 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" event={"ID":"089c681d-b6d5-43ac-9fca-ed2a92d59441","Type":"ContainerDied","Data":"5c8d5fff6a51feb41575531bc7f2c2030eaa343034944857c8509c81bdc2bdb2"} Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.531996 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.534666 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64f9c49f9b-5bjhh" podStartSLOduration=3.534647288 podStartE2EDuration="3.534647288s" podCreationTimestamp="2025-12-02 13:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:44.530357242 +0000 UTC m=+1095.486998927" watchObservedRunningTime="2025-12-02 13:22:44.534647288 +0000 UTC m=+1095.491288983" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.659825 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-784f98d458-wsm8f"] Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.763572 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b5b59474f-qnsqr"] Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.766044 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.772347 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.772720 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.791607 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b5b59474f-qnsqr"] Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922655 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-combined-ca-bundle\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922729 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-internal-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922779 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-ovndb-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922810 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-public-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922887 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-httpd-config\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922912 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-config\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:44 crc kubenswrapper[4725]: I1202 13:22:44.922978 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tng55\" (UniqueName: \"kubernetes.io/projected/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-kube-api-access-tng55\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026639 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-httpd-config\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026707 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-config\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026791 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tng55\" (UniqueName: \"kubernetes.io/projected/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-kube-api-access-tng55\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026835 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-combined-ca-bundle\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026876 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-internal-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026930 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-ovndb-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.026963 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-public-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.036674 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-config\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.039280 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-httpd-config\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.039952 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-internal-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.040616 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-public-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.042374 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-combined-ca-bundle\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.043364 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-ovndb-tls-certs\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.072444 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tng55\" (UniqueName: \"kubernetes.io/projected/e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d-kube-api-access-tng55\") pod \"neutron-7b5b59474f-qnsqr\" (UID: \"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d\") " pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.227597 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:45 crc kubenswrapper[4725]: I1202 13:22:45.567668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-784f98d458-wsm8f" event={"ID":"66a69b36-610f-4f99-b57b-834d0a8a07a1","Type":"ContainerStarted","Data":"110faf9dc805503deccdc2e500c5a14f02504b413c4c4342f306d31abb94330d"} Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.190372 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.587651 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.617807 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.617885 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.734791 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.734862 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:22:46 crc kubenswrapper[4725]: I1202 13:22:46.741484 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.489214 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b5b59474f-qnsqr"] Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.602055 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29ba5625-386c-4e5f-8aa1-7ed349af1d7a","Type":"ContainerStarted","Data":"400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e"} Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.605742 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" event={"ID":"089c681d-b6d5-43ac-9fca-ed2a92d59441","Type":"ContainerStarted","Data":"40da217e5fc67e9c37dfc1b392f5ffeaa1136f22348a1786f35addb428b40375"} Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.605948 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.609089 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-784f98d458-wsm8f" event={"ID":"66a69b36-610f-4f99-b57b-834d0a8a07a1","Type":"ContainerStarted","Data":"0e1ecebf411028a4713f14049c4d65585cb2b52c1732914fddebe46d34ee99a6"} Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.746190 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.747145 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:22:47 crc kubenswrapper[4725]: I1202 13:22:47.783248 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" podStartSLOduration=6.783232825 podStartE2EDuration="6.783232825s" podCreationTimestamp="2025-12-02 13:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:47.633953059 +0000 UTC m=+1098.590594764" watchObservedRunningTime="2025-12-02 13:22:47.783232825 +0000 UTC m=+1098.739874520" Dec 02 13:22:48 crc kubenswrapper[4725]: I1202 13:22:48.620433 4725 generic.go:334] "Generic (PLEG): container finished" podID="3128bc10-71dd-46f4-8e71-787260ede3d0" containerID="defad413a0b029b7aa4a113098e7b7a85fd86214d0d36a873dc97ae4d29e840d" exitCode=0 Dec 02 13:22:48 crc kubenswrapper[4725]: I1202 13:22:48.620499 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzkrh" event={"ID":"3128bc10-71dd-46f4-8e71-787260ede3d0","Type":"ContainerDied","Data":"defad413a0b029b7aa4a113098e7b7a85fd86214d0d36a873dc97ae4d29e840d"} Dec 02 13:22:49 crc kubenswrapper[4725]: I1202 13:22:49.128983 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 13:22:51 crc kubenswrapper[4725]: I1202 13:22:51.955292 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.039859 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fxcbw"] Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.040133 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="dnsmasq-dns" containerID="cri-o://3d89219dd82c7d2ce15dd379b4c86cf34afed0f7171061fa2878b6089deb26fa" gracePeriod=10 Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.284183 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.400063 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-config-data\") pod \"3128bc10-71dd-46f4-8e71-787260ede3d0\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.401581 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-fernet-keys\") pod \"3128bc10-71dd-46f4-8e71-787260ede3d0\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.401780 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-combined-ca-bundle\") pod \"3128bc10-71dd-46f4-8e71-787260ede3d0\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.402087 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-scripts\") pod \"3128bc10-71dd-46f4-8e71-787260ede3d0\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.402111 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-credential-keys\") pod \"3128bc10-71dd-46f4-8e71-787260ede3d0\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.404803 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hsnm\" (UniqueName: \"kubernetes.io/projected/3128bc10-71dd-46f4-8e71-787260ede3d0-kube-api-access-9hsnm\") pod \"3128bc10-71dd-46f4-8e71-787260ede3d0\" (UID: \"3128bc10-71dd-46f4-8e71-787260ede3d0\") " Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.417866 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-scripts" (OuterVolumeSpecName: "scripts") pod "3128bc10-71dd-46f4-8e71-787260ede3d0" (UID: "3128bc10-71dd-46f4-8e71-787260ede3d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.417873 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3128bc10-71dd-46f4-8e71-787260ede3d0-kube-api-access-9hsnm" (OuterVolumeSpecName: "kube-api-access-9hsnm") pod "3128bc10-71dd-46f4-8e71-787260ede3d0" (UID: "3128bc10-71dd-46f4-8e71-787260ede3d0"). InnerVolumeSpecName "kube-api-access-9hsnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.423851 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3128bc10-71dd-46f4-8e71-787260ede3d0" (UID: "3128bc10-71dd-46f4-8e71-787260ede3d0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.424891 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3128bc10-71dd-46f4-8e71-787260ede3d0" (UID: "3128bc10-71dd-46f4-8e71-787260ede3d0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.461605 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3128bc10-71dd-46f4-8e71-787260ede3d0" (UID: "3128bc10-71dd-46f4-8e71-787260ede3d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.483796 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-config-data" (OuterVolumeSpecName: "config-data") pod "3128bc10-71dd-46f4-8e71-787260ede3d0" (UID: "3128bc10-71dd-46f4-8e71-787260ede3d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.513767 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.513805 4725 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.513816 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.513829 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.513839 4725 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3128bc10-71dd-46f4-8e71-787260ede3d0-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.513850 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hsnm\" (UniqueName: \"kubernetes.io/projected/3128bc10-71dd-46f4-8e71-787260ede3d0-kube-api-access-9hsnm\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.547999 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b5b59474f-qnsqr" event={"ID":"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d","Type":"ContainerStarted","Data":"18e8b3d1ca2ca22f0bed7af4d465749f51278e7ec5452c17b3042ee4f29cdc28"} Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.549822 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kzkrh" event={"ID":"3128bc10-71dd-46f4-8e71-787260ede3d0","Type":"ContainerDied","Data":"c2db4ef7011627e7305823e5fc06fa145a1cbf465669cb3cd6c90792999b81e8"} Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.549862 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2db4ef7011627e7305823e5fc06fa145a1cbf465669cb3cd6c90792999b81e8" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.549936 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kzkrh" Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.552983 4725 generic.go:334] "Generic (PLEG): container finished" podID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerID="3d89219dd82c7d2ce15dd379b4c86cf34afed0f7171061fa2878b6089deb26fa" exitCode=0 Dec 02 13:22:52 crc kubenswrapper[4725]: I1202 13:22:52.553020 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" event={"ID":"9c262e77-097d-4bd1-aebf-dc0a401fdb5b","Type":"ContainerDied","Data":"3d89219dd82c7d2ce15dd379b4c86cf34afed0f7171061fa2878b6089deb26fa"} Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.412767 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c7dc85fc9-tj9wl"] Dec 02 13:22:53 crc kubenswrapper[4725]: E1202 13:22:53.415216 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3128bc10-71dd-46f4-8e71-787260ede3d0" containerName="keystone-bootstrap" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.415240 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="3128bc10-71dd-46f4-8e71-787260ede3d0" containerName="keystone-bootstrap" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.415422 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="3128bc10-71dd-46f4-8e71-787260ede3d0" containerName="keystone-bootstrap" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.416096 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.420754 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.420849 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.420767 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-twszr" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.421049 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.421264 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.421644 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.434000 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c7dc85fc9-tj9wl"] Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.541738 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-fernet-keys\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.541808 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-public-tls-certs\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.541903 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b26zv\" (UniqueName: \"kubernetes.io/projected/16fb266f-4c1b-406e-b662-442a92366486-kube-api-access-b26zv\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.541942 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-config-data\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.541984 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-credential-keys\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.542139 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-internal-tls-certs\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.542224 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-scripts\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.542278 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-combined-ca-bundle\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644275 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-internal-tls-certs\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644358 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-scripts\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644397 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-combined-ca-bundle\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644489 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-fernet-keys\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644512 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-public-tls-certs\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644585 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b26zv\" (UniqueName: \"kubernetes.io/projected/16fb266f-4c1b-406e-b662-442a92366486-kube-api-access-b26zv\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644631 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-config-data\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.644678 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-credential-keys\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.652443 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-credential-keys\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.653497 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-scripts\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.654868 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-config-data\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.655146 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-fernet-keys\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.655406 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-internal-tls-certs\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.655604 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-public-tls-certs\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.674201 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16fb266f-4c1b-406e-b662-442a92366486-combined-ca-bundle\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.697222 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b26zv\" (UniqueName: \"kubernetes.io/projected/16fb266f-4c1b-406e-b662-442a92366486-kube-api-access-b26zv\") pod \"keystone-6c7dc85fc9-tj9wl\" (UID: \"16fb266f-4c1b-406e-b662-442a92366486\") " pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:53 crc kubenswrapper[4725]: I1202 13:22:53.761298 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:54 crc kubenswrapper[4725]: I1202 13:22:54.284698 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:22:54 crc kubenswrapper[4725]: I1202 13:22:54.285033 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:22:56 crc kubenswrapper[4725]: I1202 13:22:56.607546 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29ba5625-386c-4e5f-8aa1-7ed349af1d7a","Type":"ContainerStarted","Data":"5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75"} Dec 02 13:22:56 crc kubenswrapper[4725]: I1202 13:22:56.621022 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 02 13:22:56 crc kubenswrapper[4725]: I1202 13:22:56.642623 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=15.642598892 podStartE2EDuration="15.642598892s" podCreationTimestamp="2025-12-02 13:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:56.627145255 +0000 UTC m=+1107.583786960" watchObservedRunningTime="2025-12-02 13:22:56.642598892 +0000 UTC m=+1107.599240587" Dec 02 13:22:56 crc kubenswrapper[4725]: I1202 13:22:56.741444 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f9f547644-22mp9" podUID="a4cfe8aa-4818-4626-8e86-6607e7bb38d2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 02 13:22:56 crc kubenswrapper[4725]: I1202 13:22:56.936761 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.014101 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-sb\") pod \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.014267 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkfvn\" (UniqueName: \"kubernetes.io/projected/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-kube-api-access-pkfvn\") pod \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.014310 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-nb\") pod \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.014417 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-svc\") pod \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.014450 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-swift-storage-0\") pod \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.014488 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-config\") pod \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\" (UID: \"9c262e77-097d-4bd1-aebf-dc0a401fdb5b\") " Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.039737 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-kube-api-access-pkfvn" (OuterVolumeSpecName: "kube-api-access-pkfvn") pod "9c262e77-097d-4bd1-aebf-dc0a401fdb5b" (UID: "9c262e77-097d-4bd1-aebf-dc0a401fdb5b"). InnerVolumeSpecName "kube-api-access-pkfvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.116631 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkfvn\" (UniqueName: \"kubernetes.io/projected/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-kube-api-access-pkfvn\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.216234 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9c262e77-097d-4bd1-aebf-dc0a401fdb5b" (UID: "9c262e77-097d-4bd1-aebf-dc0a401fdb5b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.219624 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.251994 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c7dc85fc9-tj9wl"] Dec 02 13:22:57 crc kubenswrapper[4725]: W1202 13:22:57.264576 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16fb266f_4c1b_406e_b662_442a92366486.slice/crio-a70c08dee981ec9dabcc505b8b5d39e93512d9341b462db473ca9805e9b97868 WatchSource:0}: Error finding container a70c08dee981ec9dabcc505b8b5d39e93512d9341b462db473ca9805e9b97868: Status 404 returned error can't find the container with id a70c08dee981ec9dabcc505b8b5d39e93512d9341b462db473ca9805e9b97868 Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.395203 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9c262e77-097d-4bd1-aebf-dc0a401fdb5b" (UID: "9c262e77-097d-4bd1-aebf-dc0a401fdb5b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.403910 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-config" (OuterVolumeSpecName: "config") pod "9c262e77-097d-4bd1-aebf-dc0a401fdb5b" (UID: "9c262e77-097d-4bd1-aebf-dc0a401fdb5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.405108 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9c262e77-097d-4bd1-aebf-dc0a401fdb5b" (UID: "9c262e77-097d-4bd1-aebf-dc0a401fdb5b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.415001 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9c262e77-097d-4bd1-aebf-dc0a401fdb5b" (UID: "9c262e77-097d-4bd1-aebf-dc0a401fdb5b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.432372 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.432396 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.432405 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.432415 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c262e77-097d-4bd1-aebf-dc0a401fdb5b-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.631721 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-784f98d458-wsm8f" event={"ID":"66a69b36-610f-4f99-b57b-834d0a8a07a1","Type":"ContainerStarted","Data":"e871ebcffe216a45e3be607e64ac112831c62b14138d46432ebc881f9bfe2381"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.633126 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.633519 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.637187 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" event={"ID":"9c262e77-097d-4bd1-aebf-dc0a401fdb5b","Type":"ContainerDied","Data":"29e0cbdb8a7e70bdecde52a551306a6ac1f778f2340130f605b704cd3e66d804"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.637239 4725 scope.go:117] "RemoveContainer" containerID="3d89219dd82c7d2ce15dd379b4c86cf34afed0f7171061fa2878b6089deb26fa" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.637388 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.655967 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c7dc85fc9-tj9wl" event={"ID":"16fb266f-4c1b-406e-b662-442a92366486","Type":"ContainerStarted","Data":"60f010930e46d941e84d99133743e9a97be23d1d54f1855bdee7ce69ce201523"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.656023 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.656051 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c7dc85fc9-tj9wl" event={"ID":"16fb266f-4c1b-406e-b662-442a92366486","Type":"ContainerStarted","Data":"a70c08dee981ec9dabcc505b8b5d39e93512d9341b462db473ca9805e9b97868"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.666032 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerStarted","Data":"7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.668552 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-784f98d458-wsm8f" podStartSLOduration=14.668530995 podStartE2EDuration="14.668530995s" podCreationTimestamp="2025-12-02 13:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:57.661781836 +0000 UTC m=+1108.618423531" watchObservedRunningTime="2025-12-02 13:22:57.668530995 +0000 UTC m=+1108.625172690" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.670919 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b5b59474f-qnsqr" event={"ID":"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d","Type":"ContainerStarted","Data":"f1cc4ac3580843831436cd37c52c4ab9de91e1943cfc928c950fe2f18029bb96"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.673548 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9blbh" event={"ID":"52677bda-0bda-4d69-af01-f6ce198582f6","Type":"ContainerStarted","Data":"68d371efd80ed263528e6349b6ee85336db6e7d51973a405585eb2a9f5a2e8e3"} Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.689322 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c7dc85fc9-tj9wl" podStartSLOduration=4.689308315 podStartE2EDuration="4.689308315s" podCreationTimestamp="2025-12-02 13:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:57.688557966 +0000 UTC m=+1108.645199671" watchObservedRunningTime="2025-12-02 13:22:57.689308315 +0000 UTC m=+1108.645950010" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.803776 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-9blbh" podStartSLOduration=5.075184974 podStartE2EDuration="52.803748779s" podCreationTimestamp="2025-12-02 13:22:05 +0000 UTC" firstStartedPulling="2025-12-02 13:22:08.996035099 +0000 UTC m=+1059.952676794" lastFinishedPulling="2025-12-02 13:22:56.724598904 +0000 UTC m=+1107.681240599" observedRunningTime="2025-12-02 13:22:57.713654464 +0000 UTC m=+1108.670296159" watchObservedRunningTime="2025-12-02 13:22:57.803748779 +0000 UTC m=+1108.760390474" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.804640 4725 scope.go:117] "RemoveContainer" containerID="b0bb34439059e7590dcd8a6433c94ed8a900084ea5458c0fe445eb78b85f0c99" Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.806993 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fxcbw"] Dec 02 13:22:57 crc kubenswrapper[4725]: I1202 13:22:57.904773 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fxcbw"] Dec 02 13:22:58 crc kubenswrapper[4725]: I1202 13:22:58.681240 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdwcp" event={"ID":"11048035-cfbd-4e03-8427-83e0d2a91a63","Type":"ContainerStarted","Data":"ca608bf96ccf5468c525dacc3104084d32b250d892bac9d8deb4b016b1735347"} Dec 02 13:22:58 crc kubenswrapper[4725]: I1202 13:22:58.684703 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b5b59474f-qnsqr" event={"ID":"e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d","Type":"ContainerStarted","Data":"dac768b9bfa02c45536cb87f997cbec593a7f732c4b7f41560f6560ac51e8b68"} Dec 02 13:22:58 crc kubenswrapper[4725]: I1202 13:22:58.684907 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:22:58 crc kubenswrapper[4725]: I1202 13:22:58.711352 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-pdwcp" podStartSLOduration=6.021808017 podStartE2EDuration="53.711332519s" podCreationTimestamp="2025-12-02 13:22:05 +0000 UTC" firstStartedPulling="2025-12-02 13:22:09.015517103 +0000 UTC m=+1059.972158798" lastFinishedPulling="2025-12-02 13:22:56.705041605 +0000 UTC m=+1107.661683300" observedRunningTime="2025-12-02 13:22:58.704493898 +0000 UTC m=+1109.661135603" watchObservedRunningTime="2025-12-02 13:22:58.711332519 +0000 UTC m=+1109.667974204" Dec 02 13:22:59 crc kubenswrapper[4725]: I1202 13:22:59.317073 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" path="/var/lib/kubelet/pods/9c262e77-097d-4bd1-aebf-dc0a401fdb5b/volumes" Dec 02 13:23:00 crc kubenswrapper[4725]: I1202 13:23:00.217574 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:23:00 crc kubenswrapper[4725]: I1202 13:23:00.242490 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b5b59474f-qnsqr" podStartSLOduration=16.242471144 podStartE2EDuration="16.242471144s" podCreationTimestamp="2025-12-02 13:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:22:58.758581041 +0000 UTC m=+1109.715222736" watchObservedRunningTime="2025-12-02 13:23:00.242471144 +0000 UTC m=+1111.199112839" Dec 02 13:23:01 crc kubenswrapper[4725]: I1202 13:23:01.737160 4725 generic.go:334] "Generic (PLEG): container finished" podID="52677bda-0bda-4d69-af01-f6ce198582f6" containerID="68d371efd80ed263528e6349b6ee85336db6e7d51973a405585eb2a9f5a2e8e3" exitCode=0 Dec 02 13:23:01 crc kubenswrapper[4725]: I1202 13:23:01.737208 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9blbh" event={"ID":"52677bda-0bda-4d69-af01-f6ce198582f6","Type":"ContainerDied","Data":"68d371efd80ed263528e6349b6ee85336db6e7d51973a405585eb2a9f5a2e8e3"} Dec 02 13:23:01 crc kubenswrapper[4725]: I1202 13:23:01.888682 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-fxcbw" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Dec 02 13:23:02 crc kubenswrapper[4725]: I1202 13:23:02.310767 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 13:23:02 crc kubenswrapper[4725]: I1202 13:23:02.311371 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 13:23:02 crc kubenswrapper[4725]: I1202 13:23:02.352220 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 13:23:02 crc kubenswrapper[4725]: I1202 13:23:02.362040 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 13:23:02 crc kubenswrapper[4725]: I1202 13:23:02.755021 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 13:23:02 crc kubenswrapper[4725]: I1202 13:23:02.755061 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 13:23:04 crc kubenswrapper[4725]: I1202 13:23:04.793508 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:23:04 crc kubenswrapper[4725]: I1202 13:23:04.794107 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:23:05 crc kubenswrapper[4725]: I1202 13:23:05.794585 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 13:23:05 crc kubenswrapper[4725]: I1202 13:23:05.829999 4725 generic.go:334] "Generic (PLEG): container finished" podID="11048035-cfbd-4e03-8427-83e0d2a91a63" containerID="ca608bf96ccf5468c525dacc3104084d32b250d892bac9d8deb4b016b1735347" exitCode=0 Dec 02 13:23:05 crc kubenswrapper[4725]: I1202 13:23:05.830139 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:23:05 crc kubenswrapper[4725]: I1202 13:23:05.830161 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdwcp" event={"ID":"11048035-cfbd-4e03-8427-83e0d2a91a63","Type":"ContainerDied","Data":"ca608bf96ccf5468c525dacc3104084d32b250d892bac9d8deb4b016b1735347"} Dec 02 13:23:06 crc kubenswrapper[4725]: I1202 13:23:06.225623 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 13:23:06 crc kubenswrapper[4725]: I1202 13:23:06.618798 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 02 13:23:06 crc kubenswrapper[4725]: I1202 13:23:06.729528 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f9f547644-22mp9" podUID="a4cfe8aa-4818-4626-8e86-6607e7bb38d2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.482694 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9blbh" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.498665 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527415 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-scripts\") pod \"11048035-cfbd-4e03-8427-83e0d2a91a63\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527516 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-combined-ca-bundle\") pod \"11048035-cfbd-4e03-8427-83e0d2a91a63\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527566 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9nbc\" (UniqueName: \"kubernetes.io/projected/52677bda-0bda-4d69-af01-f6ce198582f6-kube-api-access-z9nbc\") pod \"52677bda-0bda-4d69-af01-f6ce198582f6\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527635 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-config-data\") pod \"11048035-cfbd-4e03-8427-83e0d2a91a63\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527689 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-db-sync-config-data\") pod \"11048035-cfbd-4e03-8427-83e0d2a91a63\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527727 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-db-sync-config-data\") pod \"52677bda-0bda-4d69-af01-f6ce198582f6\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527774 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-combined-ca-bundle\") pod \"52677bda-0bda-4d69-af01-f6ce198582f6\" (UID: \"52677bda-0bda-4d69-af01-f6ce198582f6\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527814 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11048035-cfbd-4e03-8427-83e0d2a91a63-etc-machine-id\") pod \"11048035-cfbd-4e03-8427-83e0d2a91a63\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.527854 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96b4p\" (UniqueName: \"kubernetes.io/projected/11048035-cfbd-4e03-8427-83e0d2a91a63-kube-api-access-96b4p\") pod \"11048035-cfbd-4e03-8427-83e0d2a91a63\" (UID: \"11048035-cfbd-4e03-8427-83e0d2a91a63\") " Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.542084 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11048035-cfbd-4e03-8427-83e0d2a91a63-kube-api-access-96b4p" (OuterVolumeSpecName: "kube-api-access-96b4p") pod "11048035-cfbd-4e03-8427-83e0d2a91a63" (UID: "11048035-cfbd-4e03-8427-83e0d2a91a63"). InnerVolumeSpecName "kube-api-access-96b4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.542973 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/11048035-cfbd-4e03-8427-83e0d2a91a63-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "11048035-cfbd-4e03-8427-83e0d2a91a63" (UID: "11048035-cfbd-4e03-8427-83e0d2a91a63"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.584723 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52677bda-0bda-4d69-af01-f6ce198582f6-kube-api-access-z9nbc" (OuterVolumeSpecName: "kube-api-access-z9nbc") pod "52677bda-0bda-4d69-af01-f6ce198582f6" (UID: "52677bda-0bda-4d69-af01-f6ce198582f6"). InnerVolumeSpecName "kube-api-access-z9nbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.587489 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "11048035-cfbd-4e03-8427-83e0d2a91a63" (UID: "11048035-cfbd-4e03-8427-83e0d2a91a63"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.592552 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "52677bda-0bda-4d69-af01-f6ce198582f6" (UID: "52677bda-0bda-4d69-af01-f6ce198582f6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.607090 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-scripts" (OuterVolumeSpecName: "scripts") pod "11048035-cfbd-4e03-8427-83e0d2a91a63" (UID: "11048035-cfbd-4e03-8427-83e0d2a91a63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.629510 4725 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.629555 4725 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.629565 4725 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/11048035-cfbd-4e03-8427-83e0d2a91a63-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.629576 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96b4p\" (UniqueName: \"kubernetes.io/projected/11048035-cfbd-4e03-8427-83e0d2a91a63-kube-api-access-96b4p\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.629591 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.629600 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9nbc\" (UniqueName: \"kubernetes.io/projected/52677bda-0bda-4d69-af01-f6ce198582f6-kube-api-access-z9nbc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.650047 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11048035-cfbd-4e03-8427-83e0d2a91a63" (UID: "11048035-cfbd-4e03-8427-83e0d2a91a63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.653921 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-config-data" (OuterVolumeSpecName: "config-data") pod "11048035-cfbd-4e03-8427-83e0d2a91a63" (UID: "11048035-cfbd-4e03-8427-83e0d2a91a63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.674813 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52677bda-0bda-4d69-af01-f6ce198582f6" (UID: "52677bda-0bda-4d69-af01-f6ce198582f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.730644 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.730690 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11048035-cfbd-4e03-8427-83e0d2a91a63-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.730701 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52677bda-0bda-4d69-af01-f6ce198582f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.867893 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-9blbh" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.867885 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-9blbh" event={"ID":"52677bda-0bda-4d69-af01-f6ce198582f6","Type":"ContainerDied","Data":"1050fbb08b2302a2f38baf6c4b201214cf518b366ce2bddc683cbe2e981baf54"} Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.868403 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1050fbb08b2302a2f38baf6c4b201214cf518b366ce2bddc683cbe2e981baf54" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.871814 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-pdwcp" event={"ID":"11048035-cfbd-4e03-8427-83e0d2a91a63","Type":"ContainerDied","Data":"e908e15b5237bd932306de557c2e39a553aa6f06c0b7c9afc7badd405b6bea8d"} Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.871961 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e908e15b5237bd932306de557c2e39a553aa6f06c0b7c9afc7badd405b6bea8d" Dec 02 13:23:08 crc kubenswrapper[4725]: I1202 13:23:08.872116 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-pdwcp" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.770014 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.868181 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rtmv\" (UniqueName: \"kubernetes.io/projected/df2ed35d-e2f2-4c5f-b2db-12d95291d167-kube-api-access-9rtmv\") pod \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.868271 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df2ed35d-e2f2-4c5f-b2db-12d95291d167-horizon-secret-key\") pod \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.868367 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-config-data\") pod \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.868406 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-scripts\") pod \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.868446 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df2ed35d-e2f2-4c5f-b2db-12d95291d167-logs\") pod \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\" (UID: \"df2ed35d-e2f2-4c5f-b2db-12d95291d167\") " Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.869592 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df2ed35d-e2f2-4c5f-b2db-12d95291d167-logs" (OuterVolumeSpecName: "logs") pod "df2ed35d-e2f2-4c5f-b2db-12d95291d167" (UID: "df2ed35d-e2f2-4c5f-b2db-12d95291d167"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.901315 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df2ed35d-e2f2-4c5f-b2db-12d95291d167-kube-api-access-9rtmv" (OuterVolumeSpecName: "kube-api-access-9rtmv") pod "df2ed35d-e2f2-4c5f-b2db-12d95291d167" (UID: "df2ed35d-e2f2-4c5f-b2db-12d95291d167"). InnerVolumeSpecName "kube-api-access-9rtmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.923365 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df2ed35d-e2f2-4c5f-b2db-12d95291d167-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "df2ed35d-e2f2-4c5f-b2db-12d95291d167" (UID: "df2ed35d-e2f2-4c5f-b2db-12d95291d167"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.951809 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.952417 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon-log" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952435 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon-log" Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.952467 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="init" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952474 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="init" Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.952492 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="dnsmasq-dns" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952498 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="dnsmasq-dns" Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.952507 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952523 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon" Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.952541 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11048035-cfbd-4e03-8427-83e0d2a91a63" containerName="cinder-db-sync" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952547 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="11048035-cfbd-4e03-8427-83e0d2a91a63" containerName="cinder-db-sync" Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.952565 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52677bda-0bda-4d69-af01-f6ce198582f6" containerName="barbican-db-sync" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952571 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="52677bda-0bda-4d69-af01-f6ce198582f6" containerName="barbican-db-sync" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952802 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952824 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerName="horizon-log" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952850 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="52677bda-0bda-4d69-af01-f6ce198582f6" containerName="barbican-db-sync" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952859 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c262e77-097d-4bd1-aebf-dc0a401fdb5b" containerName="dnsmasq-dns" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.952870 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="11048035-cfbd-4e03-8427-83e0d2a91a63" containerName="cinder-db-sync" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.954149 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.967299 4725 generic.go:334] "Generic (PLEG): container finished" podID="5352a37f-00de-4087-8337-6536a0c03fd5" containerID="c2a0785f7ce3393a1b5e7ddbe3ba5ef2b58150b80527a0ce28d3f98863104a5d" exitCode=137 Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.967337 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zd7gg" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.967440 4725 generic.go:334] "Generic (PLEG): container finished" podID="5352a37f-00de-4087-8337-6536a0c03fd5" containerID="1e73159d6ea8897fc75b41f9b67e58dc11bc565d76fa18bd6fad3abfa225b947" exitCode=137 Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.967836 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.968014 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.974375 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.974552 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwghx\" (UniqueName: \"kubernetes.io/projected/1fe1f66d-914d-4e77-9d98-e55681fdc408-kube-api-access-gwghx\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.974732 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.974903 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.974946 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fe1f66d-914d-4e77-9d98-e55681fdc408-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.974974 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-scripts\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.975443 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rtmv\" (UniqueName: \"kubernetes.io/projected/df2ed35d-e2f2-4c5f-b2db-12d95291d167-kube-api-access-9rtmv\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.975488 4725 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/df2ed35d-e2f2-4c5f-b2db-12d95291d167-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.975500 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df2ed35d-e2f2-4c5f-b2db-12d95291d167-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.979903 4725 generic.go:334] "Generic (PLEG): container finished" podID="de526deb-a82c-469a-9387-308555f0d667" containerID="f8e3cd60233f0e0901e1480751dc5a6c7379feeb0fe885446b85d4ff1b41ff30" exitCode=137 Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.979989 4725 generic.go:334] "Generic (PLEG): container finished" podID="de526deb-a82c-469a-9387-308555f0d667" containerID="73f8e7a7c9725ec6b7f049eec81a1ae192748e5d7b3c158a640ccbf9c18d217f" exitCode=137 Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.980641 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.967515 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77f7bfffdc-cnbqk" event={"ID":"5352a37f-00de-4087-8337-6536a0c03fd5","Type":"ContainerDied","Data":"c2a0785f7ce3393a1b5e7ddbe3ba5ef2b58150b80527a0ce28d3f98863104a5d"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.981411 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77f7bfffdc-cnbqk" event={"ID":"5352a37f-00de-4087-8337-6536a0c03fd5","Type":"ContainerDied","Data":"1e73159d6ea8897fc75b41f9b67e58dc11bc565d76fa18bd6fad3abfa225b947"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.981437 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5489f8c44c-g22fg" event={"ID":"de526deb-a82c-469a-9387-308555f0d667","Type":"ContainerDied","Data":"f8e3cd60233f0e0901e1480751dc5a6c7379feeb0fe885446b85d4ff1b41ff30"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.981478 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5489f8c44c-g22fg" event={"ID":"de526deb-a82c-469a-9387-308555f0d667","Type":"ContainerDied","Data":"73f8e7a7c9725ec6b7f049eec81a1ae192748e5d7b3c158a640ccbf9c18d217f"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.990638 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.991798 4725 generic.go:334] "Generic (PLEG): container finished" podID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerID="b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373" exitCode=137 Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.991866 4725 generic.go:334] "Generic (PLEG): container finished" podID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" containerID="0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b" exitCode=137 Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.991920 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f4bf4679-pfzbz" event={"ID":"df2ed35d-e2f2-4c5f-b2db-12d95291d167","Type":"ContainerDied","Data":"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.991980 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f4bf4679-pfzbz" event={"ID":"df2ed35d-e2f2-4c5f-b2db-12d95291d167","Type":"ContainerDied","Data":"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.991997 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74f4bf4679-pfzbz" event={"ID":"df2ed35d-e2f2-4c5f-b2db-12d95291d167","Type":"ContainerDied","Data":"8fcc5da26dcda3cd55b724aef41257433aae3101c0020fceb510f16c12a8c2f9"} Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.992021 4725 scope.go:117] "RemoveContainer" containerID="b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.992318 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74f4bf4679-pfzbz" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.992723 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-scripts" (OuterVolumeSpecName: "scripts") pod "df2ed35d-e2f2-4c5f-b2db-12d95291d167" (UID: "df2ed35d-e2f2-4c5f-b2db-12d95291d167"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.997140 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5c5f8cd4f5-fd8q7"] Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.997670 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.997687 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon" Dec 02 13:23:09 crc kubenswrapper[4725]: E1202 13:23:09.997728 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon-log" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.997735 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon-log" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.997961 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.997978 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" containerName="horizon-log" Dec 02 13:23:09 crc kubenswrapper[4725]: I1202 13:23:09.999112 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.016641 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-config-data" (OuterVolumeSpecName: "config-data") pod "df2ed35d-e2f2-4c5f-b2db-12d95291d167" (UID: "df2ed35d-e2f2-4c5f-b2db-12d95291d167"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.040165 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-6h2cf" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.040662 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.041568 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.088152 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-scripts\") pod \"5352a37f-00de-4087-8337-6536a0c03fd5\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.092224 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5352a37f-00de-4087-8337-6536a0c03fd5-logs\") pod \"5352a37f-00de-4087-8337-6536a0c03fd5\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.092414 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n46wz\" (UniqueName: \"kubernetes.io/projected/5352a37f-00de-4087-8337-6536a0c03fd5-kube-api-access-n46wz\") pod \"5352a37f-00de-4087-8337-6536a0c03fd5\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.092525 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5352a37f-00de-4087-8337-6536a0c03fd5-horizon-secret-key\") pod \"5352a37f-00de-4087-8337-6536a0c03fd5\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.092655 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-config-data\") pod \"5352a37f-00de-4087-8337-6536a0c03fd5\" (UID: \"5352a37f-00de-4087-8337-6536a0c03fd5\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.093347 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5352a37f-00de-4087-8337-6536a0c03fd5-logs" (OuterVolumeSpecName: "logs") pod "5352a37f-00de-4087-8337-6536a0c03fd5" (UID: "5352a37f-00de-4087-8337-6536a0c03fd5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.093784 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-config-data\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.093908 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwghx\" (UniqueName: \"kubernetes.io/projected/1fe1f66d-914d-4e77-9d98-e55681fdc408-kube-api-access-gwghx\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094113 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-config-data-custom\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094338 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-combined-ca-bundle\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094415 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643b3918-6dab-4956-bd36-474ba0195d15-logs\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094667 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094741 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fe1f66d-914d-4e77-9d98-e55681fdc408-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.094813 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-scripts\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.095096 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sbjc\" (UniqueName: \"kubernetes.io/projected/643b3918-6dab-4956-bd36-474ba0195d15-kube-api-access-9sbjc\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.095183 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.095338 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.095373 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df2ed35d-e2f2-4c5f-b2db-12d95291d167-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.095389 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5352a37f-00de-4087-8337-6536a0c03fd5-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.130380 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: E1202 13:23:10.134832 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.152626 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5352a37f-00de-4087-8337-6536a0c03fd5-kube-api-access-n46wz" (OuterVolumeSpecName: "kube-api-access-n46wz") pod "5352a37f-00de-4087-8337-6536a0c03fd5" (UID: "5352a37f-00de-4087-8337-6536a0c03fd5"). InnerVolumeSpecName "kube-api-access-n46wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.152692 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.159656 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fe1f66d-914d-4e77-9d98-e55681fdc408-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.162239 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c5f8cd4f5-fd8q7"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.175267 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5352a37f-00de-4087-8337-6536a0c03fd5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5352a37f-00de-4087-8337-6536a0c03fd5" (UID: "5352a37f-00de-4087-8337-6536a0c03fd5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.181897 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.184451 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.201808 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-54b4d7cbbb-hdq62"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.206389 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.210654 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sbjc\" (UniqueName: \"kubernetes.io/projected/643b3918-6dab-4956-bd36-474ba0195d15-kube-api-access-9sbjc\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.210751 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-config-data\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.210859 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-config-data-custom\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.210932 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-combined-ca-bundle\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.210985 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643b3918-6dab-4956-bd36-474ba0195d15-logs\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.211164 4725 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5352a37f-00de-4087-8337-6536a0c03fd5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.211179 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n46wz\" (UniqueName: \"kubernetes.io/projected/5352a37f-00de-4087-8337-6536a0c03fd5-kube-api-access-n46wz\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.213109 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.214266 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643b3918-6dab-4956-bd36-474ba0195d15-logs\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.284988 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwghx\" (UniqueName: \"kubernetes.io/projected/1fe1f66d-914d-4e77-9d98-e55681fdc408-kube-api-access-gwghx\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.286847 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-scripts\") pod \"cinder-scheduler-0\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.287354 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-combined-ca-bundle\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.287854 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-config-data\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.315652 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-config-data-custom\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.317192 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-config-data\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.317243 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2ea09-3acf-4c71-8301-c6531e608bf1-logs\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.317404 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbvlw\" (UniqueName: \"kubernetes.io/projected/e5f2ea09-3acf-4c71-8301-c6531e608bf1-kube-api-access-rbvlw\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.317443 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-combined-ca-bundle\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.317881 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-s8z76"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.326831 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.335383 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/643b3918-6dab-4956-bd36-474ba0195d15-config-data-custom\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.346475 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-scripts" (OuterVolumeSpecName: "scripts") pod "5352a37f-00de-4087-8337-6536a0c03fd5" (UID: "5352a37f-00de-4087-8337-6536a0c03fd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.349673 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54b4d7cbbb-hdq62"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.351079 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sbjc\" (UniqueName: \"kubernetes.io/projected/643b3918-6dab-4956-bd36-474ba0195d15-kube-api-access-9sbjc\") pod \"barbican-worker-5c5f8cd4f5-fd8q7\" (UID: \"643b3918-6dab-4956-bd36-474ba0195d15\") " pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.369877 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-s8z76"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.380293 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-config-data" (OuterVolumeSpecName: "config-data") pod "5352a37f-00de-4087-8337-6536a0c03fd5" (UID: "5352a37f-00de-4087-8337-6536a0c03fd5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.418071 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-s8z76"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426340 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-config\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426408 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426474 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfrmc\" (UniqueName: \"kubernetes.io/projected/cc552b8a-fee3-4999-a174-c6505a51a9fd-kube-api-access-nfrmc\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426548 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbvlw\" (UniqueName: \"kubernetes.io/projected/e5f2ea09-3acf-4c71-8301-c6531e608bf1-kube-api-access-rbvlw\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426612 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-combined-ca-bundle\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426656 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-config-data-custom\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426710 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-config-data\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426752 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426778 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426812 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2ea09-3acf-4c71-8301-c6531e608bf1-logs\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426882 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-svc\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426971 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.426986 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5352a37f-00de-4087-8337-6536a0c03fd5-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.433085 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5f2ea09-3acf-4c71-8301-c6531e608bf1-logs\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.433821 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.437648 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6644cd5fd-44gdx"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.439474 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.443216 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.449902 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-config-data-custom\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.454546 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-combined-ca-bundle\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.455106 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f2ea09-3acf-4c71-8301-c6531e608bf1-config-data\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.473507 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.494928 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-xtptm"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.498123 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.513212 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbvlw\" (UniqueName: \"kubernetes.io/projected/e5f2ea09-3acf-4c71-8301-c6531e608bf1-kube-api-access-rbvlw\") pod \"barbican-keystone-listener-54b4d7cbbb-hdq62\" (UID: \"e5f2ea09-3acf-4c71-8301-c6531e608bf1\") " pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534723 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534777 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534802 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534831 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-logs\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534860 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-svc\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534935 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-config\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534967 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.534997 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data-custom\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.535027 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfrmc\" (UniqueName: \"kubernetes.io/projected/cc552b8a-fee3-4999-a174-c6505a51a9fd-kube-api-access-nfrmc\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.535060 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-combined-ca-bundle\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.535086 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6db82\" (UniqueName: \"kubernetes.io/projected/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-kube-api-access-6db82\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.538361 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.538571 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.539276 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-svc\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.539694 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-config\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.542445 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.573051 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6644cd5fd-44gdx"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.601753 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfrmc\" (UniqueName: \"kubernetes.io/projected/cc552b8a-fee3-4999-a174-c6505a51a9fd-kube-api-access-nfrmc\") pod \"dnsmasq-dns-85ff748b95-s8z76\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.605184 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-xtptm"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.625048 4725 scope.go:117] "RemoveContainer" containerID="0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.636106 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.639787 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.639859 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.639893 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data-custom\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.639929 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.639951 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-combined-ca-bundle\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.639982 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6db82\" (UniqueName: \"kubernetes.io/projected/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-kube-api-access-6db82\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.640016 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j6jv\" (UniqueName: \"kubernetes.io/projected/0e621595-2643-4b38-913c-23bc8f80fb62-kube-api-access-2j6jv\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.640060 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.640086 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-logs\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.640114 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.640151 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-config\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.641981 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-logs\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.648419 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data-custom\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.649904 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.650324 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-combined-ca-bundle\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.652733 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.656884 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.664829 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.675580 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.682104 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6db82\" (UniqueName: \"kubernetes.io/projected/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-kube-api-access-6db82\") pod \"barbican-api-6644cd5fd-44gdx\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.707861 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.710230 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756096 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756151 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16749490-6ff5-4ffc-bd23-d47ec3082d94-etc-machine-id\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756177 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-scripts\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756203 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756248 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data-custom\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756298 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756325 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.756368 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16749490-6ff5-4ffc-bd23-d47ec3082d94-logs\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.761859 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.762382 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.769019 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j6jv\" (UniqueName: \"kubernetes.io/projected/0e621595-2643-4b38-913c-23bc8f80fb62-kube-api-access-2j6jv\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.769104 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msxlv\" (UniqueName: \"kubernetes.io/projected/16749490-6ff5-4ffc-bd23-d47ec3082d94-kube-api-access-msxlv\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.769171 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.769272 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.769294 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-config\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.770322 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-config\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.770441 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.771507 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.780529 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.792996 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j6jv\" (UniqueName: \"kubernetes.io/projected/0e621595-2643-4b38-913c-23bc8f80fb62-kube-api-access-2j6jv\") pod \"dnsmasq-dns-5c9776ccc5-xtptm\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.822834 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74f4bf4679-pfzbz"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.853413 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74f4bf4679-pfzbz"] Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.871545 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/de526deb-a82c-469a-9387-308555f0d667-horizon-secret-key\") pod \"de526deb-a82c-469a-9387-308555f0d667\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872000 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-config-data\") pod \"de526deb-a82c-469a-9387-308555f0d667\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872069 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de526deb-a82c-469a-9387-308555f0d667-logs\") pod \"de526deb-a82c-469a-9387-308555f0d667\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872165 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-scripts\") pod \"de526deb-a82c-469a-9387-308555f0d667\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872204 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9zf7\" (UniqueName: \"kubernetes.io/projected/de526deb-a82c-469a-9387-308555f0d667-kube-api-access-m9zf7\") pod \"de526deb-a82c-469a-9387-308555f0d667\" (UID: \"de526deb-a82c-469a-9387-308555f0d667\") " Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872453 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de526deb-a82c-469a-9387-308555f0d667-logs" (OuterVolumeSpecName: "logs") pod "de526deb-a82c-469a-9387-308555f0d667" (UID: "de526deb-a82c-469a-9387-308555f0d667"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872686 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16749490-6ff5-4ffc-bd23-d47ec3082d94-etc-machine-id\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872719 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-scripts\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872855 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data-custom\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.872880 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.873017 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16749490-6ff5-4ffc-bd23-d47ec3082d94-logs\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.873088 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msxlv\" (UniqueName: \"kubernetes.io/projected/16749490-6ff5-4ffc-bd23-d47ec3082d94-kube-api-access-msxlv\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.873151 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.873316 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de526deb-a82c-469a-9387-308555f0d667-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.874040 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16749490-6ff5-4ffc-bd23-d47ec3082d94-logs\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.874144 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16749490-6ff5-4ffc-bd23-d47ec3082d94-etc-machine-id\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.900824 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.906273 4725 scope.go:117] "RemoveContainer" containerID="b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373" Dec 02 13:23:10 crc kubenswrapper[4725]: E1202 13:23:10.906741 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373\": container with ID starting with b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373 not found: ID does not exist" containerID="b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.906981 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373"} err="failed to get container status \"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373\": rpc error: code = NotFound desc = could not find container \"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373\": container with ID starting with b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373 not found: ID does not exist" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.907008 4725 scope.go:117] "RemoveContainer" containerID="0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b" Dec 02 13:23:10 crc kubenswrapper[4725]: E1202 13:23:10.907614 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b\": container with ID starting with 0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b not found: ID does not exist" containerID="0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.907639 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b"} err="failed to get container status \"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b\": rpc error: code = NotFound desc = could not find container \"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b\": container with ID starting with 0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b not found: ID does not exist" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.907687 4725 scope.go:117] "RemoveContainer" containerID="b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.908520 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373"} err="failed to get container status \"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373\": rpc error: code = NotFound desc = could not find container \"b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373\": container with ID starting with b62ca507280d77c979c87e0ff165ff0eee0f17056afa6a9ffe2a935acfbcb373 not found: ID does not exist" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.908543 4725 scope.go:117] "RemoveContainer" containerID="0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.909229 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b"} err="failed to get container status \"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b\": rpc error: code = NotFound desc = could not find container \"0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b\": container with ID starting with 0981dad5ced3811fd3a32fcd75a00a139c991d957764b2995ec8d69886ca4b7b not found: ID does not exist" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.912316 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de526deb-a82c-469a-9387-308555f0d667-kube-api-access-m9zf7" (OuterVolumeSpecName: "kube-api-access-m9zf7") pod "de526deb-a82c-469a-9387-308555f0d667" (UID: "de526deb-a82c-469a-9387-308555f0d667"). InnerVolumeSpecName "kube-api-access-m9zf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.918127 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msxlv\" (UniqueName: \"kubernetes.io/projected/16749490-6ff5-4ffc-bd23-d47ec3082d94-kube-api-access-msxlv\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.920402 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.925075 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-scripts\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.940497 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data-custom\") pod \"cinder-api-0\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " pod="openstack/cinder-api-0" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.940864 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de526deb-a82c-469a-9387-308555f0d667-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "de526deb-a82c-469a-9387-308555f0d667" (UID: "de526deb-a82c-469a-9387-308555f0d667"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.972449 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-scripts" (OuterVolumeSpecName: "scripts") pod "de526deb-a82c-469a-9387-308555f0d667" (UID: "de526deb-a82c-469a-9387-308555f0d667"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.980000 4725 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/de526deb-a82c-469a-9387-308555f0d667-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.980032 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.980044 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9zf7\" (UniqueName: \"kubernetes.io/projected/de526deb-a82c-469a-9387-308555f0d667-kube-api-access-m9zf7\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:10 crc kubenswrapper[4725]: I1202 13:23:10.981096 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-config-data" (OuterVolumeSpecName: "config-data") pod "de526deb-a82c-469a-9387-308555f0d667" (UID: "de526deb-a82c-469a-9387-308555f0d667"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.020102 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77f7bfffdc-cnbqk" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.020108 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77f7bfffdc-cnbqk" event={"ID":"5352a37f-00de-4087-8337-6536a0c03fd5","Type":"ContainerDied","Data":"9d961ffdbca1348bf88e054d0fe9890b289b17b4edb6a1b107a5f4aadb948695"} Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.020228 4725 scope.go:117] "RemoveContainer" containerID="c2a0785f7ce3393a1b5e7ddbe3ba5ef2b58150b80527a0ce28d3f98863104a5d" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.026012 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5489f8c44c-g22fg" event={"ID":"de526deb-a82c-469a-9387-308555f0d667","Type":"ContainerDied","Data":"064d9a388db5e9c5434402b57a424de1e6fde35360bb319507863b9c5bafc9eb"} Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.026122 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5489f8c44c-g22fg" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.034167 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.040383 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerStarted","Data":"b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c"} Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.040591 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="ceilometer-notification-agent" containerID="cri-o://2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc" gracePeriod=30 Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.040939 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.041026 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="proxy-httpd" containerID="cri-o://b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c" gracePeriod=30 Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.041073 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="sg-core" containerID="cri-o://7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a" gracePeriod=30 Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.079771 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77f7bfffdc-cnbqk"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.083573 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.085021 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de526deb-a82c-469a-9387-308555f0d667-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.125371 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77f7bfffdc-cnbqk"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.178395 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5489f8c44c-g22fg"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.212879 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5489f8c44c-g22fg"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.302745 4725 scope.go:117] "RemoveContainer" containerID="1e73159d6ea8897fc75b41f9b67e58dc11bc565d76fa18bd6fad3abfa225b947" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.307031 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5352a37f-00de-4087-8337-6536a0c03fd5" path="/var/lib/kubelet/pods/5352a37f-00de-4087-8337-6536a0c03fd5/volumes" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.308664 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de526deb-a82c-469a-9387-308555f0d667" path="/var/lib/kubelet/pods/de526deb-a82c-469a-9387-308555f0d667/volumes" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.313319 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df2ed35d-e2f2-4c5f-b2db-12d95291d167" path="/var/lib/kubelet/pods/df2ed35d-e2f2-4c5f-b2db-12d95291d167/volumes" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.385912 4725 scope.go:117] "RemoveContainer" containerID="f8e3cd60233f0e0901e1480751dc5a6c7379feeb0fe885446b85d4ff1b41ff30" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.431251 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5c5f8cd4f5-fd8q7"] Dec 02 13:23:11 crc kubenswrapper[4725]: W1202 13:23:11.432827 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod643b3918_6dab_4956_bd36_474ba0195d15.slice/crio-a3e7adf5fab32d8f37a8f5a598c5eb4e9d89ed35670e5aabb9217c6b8339e479 WatchSource:0}: Error finding container a3e7adf5fab32d8f37a8f5a598c5eb4e9d89ed35670e5aabb9217c6b8339e479: Status 404 returned error can't find the container with id a3e7adf5fab32d8f37a8f5a598c5eb4e9d89ed35670e5aabb9217c6b8339e479 Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.602540 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6644cd5fd-44gdx"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.623409 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.645952 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54b4d7cbbb-hdq62"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.770722 4725 scope.go:117] "RemoveContainer" containerID="73f8e7a7c9725ec6b7f049eec81a1ae192748e5d7b3c158a640ccbf9c18d217f" Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.781380 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-s8z76"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.959742 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:11 crc kubenswrapper[4725]: I1202 13:23:11.984044 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.008171 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-xtptm"] Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.080103 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" event={"ID":"e5f2ea09-3acf-4c71-8301-c6531e608bf1","Type":"ContainerStarted","Data":"b563cf60b26126dbe6741b495e59f0c4273e4c8903d8f0b8ecee0f5e9f805a40"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.083649 4725 generic.go:334] "Generic (PLEG): container finished" podID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerID="b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c" exitCode=0 Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.083739 4725 generic.go:334] "Generic (PLEG): container finished" podID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerID="7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a" exitCode=2 Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.083696 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerDied","Data":"b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.083862 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerDied","Data":"7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.099079 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fe1f66d-914d-4e77-9d98-e55681fdc408","Type":"ContainerStarted","Data":"f95c3f67c3f6971d44d736a9b67e0538471772a3e021193d6c92195984d435fe"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.106016 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" event={"ID":"643b3918-6dab-4956-bd36-474ba0195d15","Type":"ContainerStarted","Data":"a3e7adf5fab32d8f37a8f5a598c5eb4e9d89ed35670e5aabb9217c6b8339e479"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.110083 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" event={"ID":"cc552b8a-fee3-4999-a174-c6505a51a9fd","Type":"ContainerStarted","Data":"4b2da6bca28aaa64dd64b54465dd200bc100f724610a7f59f5bb93b6e88e363f"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.117126 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"16749490-6ff5-4ffc-bd23-d47ec3082d94","Type":"ContainerStarted","Data":"118f676d2d63f8f5bf1539e85d35ca63a2970ea018f6a797ed4c82e506daf731"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.124065 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6644cd5fd-44gdx" event={"ID":"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948","Type":"ContainerStarted","Data":"6630ad5d73defc990f8549a21033b72ea9b721a8c012a582c40608896569d400"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.125668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" event={"ID":"0e621595-2643-4b38-913c-23bc8f80fb62","Type":"ContainerStarted","Data":"b84c2df8360ffe39f8ecd6030be8ad018cb68e904bafedb2e739d5c318a269ac"} Dec 02 13:23:12 crc kubenswrapper[4725]: I1202 13:23:12.510148 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.037610 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.146132 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6644cd5fd-44gdx" event={"ID":"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948","Type":"ContainerStarted","Data":"05e658ba9914f6b64a7afc3562df07ebc1c68f21d618f6a545cac5c653e83aba"} Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.146193 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6644cd5fd-44gdx" event={"ID":"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948","Type":"ContainerStarted","Data":"ac34e06a59152d1275189c8bd1313e1f87feffac653b3ac6a7ccb2f3aac0a747"} Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.148051 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.148087 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.152138 4725 generic.go:334] "Generic (PLEG): container finished" podID="0e621595-2643-4b38-913c-23bc8f80fb62" containerID="c1f1baa0e09625e9962a5936162f32f51cd35eb41b7296e241cc5250d3f005b0" exitCode=0 Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.152214 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" event={"ID":"0e621595-2643-4b38-913c-23bc8f80fb62","Type":"ContainerDied","Data":"c1f1baa0e09625e9962a5936162f32f51cd35eb41b7296e241cc5250d3f005b0"} Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.170707 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-log-httpd\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.173325 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-sg-core-conf-yaml\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.173519 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-combined-ca-bundle\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.173612 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5r6w\" (UniqueName: \"kubernetes.io/projected/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-kube-api-access-g5r6w\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.173696 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.174812 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-config-data\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.174898 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-run-httpd\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.174959 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-scripts\") pod \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\" (UID: \"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5\") " Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.176198 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.177381 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.186864 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6644cd5fd-44gdx" podStartSLOduration=3.186817618 podStartE2EDuration="3.186817618s" podCreationTimestamp="2025-12-02 13:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:13.178657574 +0000 UTC m=+1124.135299289" watchObservedRunningTime="2025-12-02 13:23:13.186817618 +0000 UTC m=+1124.143459313" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.187946 4725 generic.go:334] "Generic (PLEG): container finished" podID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerID="2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc" exitCode=0 Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.188060 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerDied","Data":"2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc"} Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.188127 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"335fa14a-e2ec-4f9e-a56d-ab2a9be708f5","Type":"ContainerDied","Data":"98355c053203c3bcb96869dbcbccd637a1baf8b8bc9356eb1ab99898b44abe87"} Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.188184 4725 scope.go:117] "RemoveContainer" containerID="b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.188314 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.192936 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-scripts" (OuterVolumeSpecName: "scripts") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.214978 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-kube-api-access-g5r6w" (OuterVolumeSpecName: "kube-api-access-g5r6w") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "kube-api-access-g5r6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.216489 4725 generic.go:334] "Generic (PLEG): container finished" podID="cc552b8a-fee3-4999-a174-c6505a51a9fd" containerID="46b16cc4c7d0e763786e97f65653f835629b8e641b67b0a6af1cb404505463d7" exitCode=0 Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.216656 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" event={"ID":"cc552b8a-fee3-4999-a174-c6505a51a9fd","Type":"ContainerDied","Data":"46b16cc4c7d0e763786e97f65653f835629b8e641b67b0a6af1cb404505463d7"} Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.268364 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.285914 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.285958 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.285969 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.285980 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5r6w\" (UniqueName: \"kubernetes.io/projected/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-kube-api-access-g5r6w\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.314742 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.359442 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-config-data" (OuterVolumeSpecName: "config-data") pod "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" (UID: "335fa14a-e2ec-4f9e-a56d-ab2a9be708f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.388665 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.388711 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.603106 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.628270 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.636121 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:13 crc kubenswrapper[4725]: E1202 13:23:13.636784 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="sg-core" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.636809 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="sg-core" Dec 02 13:23:13 crc kubenswrapper[4725]: E1202 13:23:13.636819 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.636825 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon" Dec 02 13:23:13 crc kubenswrapper[4725]: E1202 13:23:13.636961 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon-log" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.636975 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon-log" Dec 02 13:23:13 crc kubenswrapper[4725]: E1202 13:23:13.636997 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="proxy-httpd" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637006 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="proxy-httpd" Dec 02 13:23:13 crc kubenswrapper[4725]: E1202 13:23:13.637020 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="ceilometer-notification-agent" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637028 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="ceilometer-notification-agent" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637257 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon-log" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637274 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="sg-core" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637288 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="proxy-httpd" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637299 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="de526deb-a82c-469a-9387-308555f0d667" containerName="horizon" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.637314 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" containerName="ceilometer-notification-agent" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.639710 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.642012 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.642552 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.644857 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699439 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-log-httpd\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699534 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-config-data\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699563 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699606 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nhvj\" (UniqueName: \"kubernetes.io/projected/974e23a0-ac9d-419d-825e-b707b3808571-kube-api-access-2nhvj\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699679 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699718 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-run-httpd\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.699771 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-scripts\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.804619 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nhvj\" (UniqueName: \"kubernetes.io/projected/974e23a0-ac9d-419d-825e-b707b3808571-kube-api-access-2nhvj\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.806305 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.806372 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-run-httpd\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.808715 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-scripts\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.809009 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-log-httpd\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.809086 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-config-data\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.809114 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.811130 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-run-httpd\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.811274 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-log-httpd\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.816195 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.818264 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.822971 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-scripts\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.823996 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-config-data\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.831669 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nhvj\" (UniqueName: \"kubernetes.io/projected/974e23a0-ac9d-419d-825e-b707b3808571-kube-api-access-2nhvj\") pod \"ceilometer-0\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " pod="openstack/ceilometer-0" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.869320 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-784f98d458-wsm8f" Dec 02 13:23:13 crc kubenswrapper[4725]: I1202 13:23:13.968340 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.069746 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.117755 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-nb\") pod \"cc552b8a-fee3-4999-a174-c6505a51a9fd\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.117832 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfrmc\" (UniqueName: \"kubernetes.io/projected/cc552b8a-fee3-4999-a174-c6505a51a9fd-kube-api-access-nfrmc\") pod \"cc552b8a-fee3-4999-a174-c6505a51a9fd\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.117956 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-swift-storage-0\") pod \"cc552b8a-fee3-4999-a174-c6505a51a9fd\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.117984 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-sb\") pod \"cc552b8a-fee3-4999-a174-c6505a51a9fd\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.118012 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-svc\") pod \"cc552b8a-fee3-4999-a174-c6505a51a9fd\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.118097 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-config\") pod \"cc552b8a-fee3-4999-a174-c6505a51a9fd\" (UID: \"cc552b8a-fee3-4999-a174-c6505a51a9fd\") " Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.124355 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc552b8a-fee3-4999-a174-c6505a51a9fd-kube-api-access-nfrmc" (OuterVolumeSpecName: "kube-api-access-nfrmc") pod "cc552b8a-fee3-4999-a174-c6505a51a9fd" (UID: "cc552b8a-fee3-4999-a174-c6505a51a9fd"). InnerVolumeSpecName "kube-api-access-nfrmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.150233 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-config" (OuterVolumeSpecName: "config") pod "cc552b8a-fee3-4999-a174-c6505a51a9fd" (UID: "cc552b8a-fee3-4999-a174-c6505a51a9fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.165283 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cc552b8a-fee3-4999-a174-c6505a51a9fd" (UID: "cc552b8a-fee3-4999-a174-c6505a51a9fd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.167993 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cc552b8a-fee3-4999-a174-c6505a51a9fd" (UID: "cc552b8a-fee3-4999-a174-c6505a51a9fd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.174305 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cc552b8a-fee3-4999-a174-c6505a51a9fd" (UID: "cc552b8a-fee3-4999-a174-c6505a51a9fd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.175344 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cc552b8a-fee3-4999-a174-c6505a51a9fd" (UID: "cc552b8a-fee3-4999-a174-c6505a51a9fd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.220594 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfrmc\" (UniqueName: \"kubernetes.io/projected/cc552b8a-fee3-4999-a174-c6505a51a9fd-kube-api-access-nfrmc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.220631 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.220641 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.220652 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.220662 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.220673 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cc552b8a-fee3-4999-a174-c6505a51a9fd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.254782 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"16749490-6ff5-4ffc-bd23-d47ec3082d94","Type":"ContainerStarted","Data":"c984a821133bc14a4ba4cad372309cb11a7e2488de1b4d8046e42c452c9ab432"} Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.274683 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.282118 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-s8z76" event={"ID":"cc552b8a-fee3-4999-a174-c6505a51a9fd","Type":"ContainerDied","Data":"4b2da6bca28aaa64dd64b54465dd200bc100f724610a7f59f5bb93b6e88e363f"} Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.349882 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-s8z76"] Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.367213 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-s8z76"] Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.739041 4725 scope.go:117] "RemoveContainer" containerID="7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.801422 4725 scope.go:117] "RemoveContainer" containerID="2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.946101 4725 scope.go:117] "RemoveContainer" containerID="b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c" Dec 02 13:23:14 crc kubenswrapper[4725]: E1202 13:23:14.946760 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c\": container with ID starting with b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c not found: ID does not exist" containerID="b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.946804 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c"} err="failed to get container status \"b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c\": rpc error: code = NotFound desc = could not find container \"b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c\": container with ID starting with b2f54368e317eaf4e1e9b12797540a261cc5a77093fea6af4fed736f04ef0a8c not found: ID does not exist" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.946832 4725 scope.go:117] "RemoveContainer" containerID="7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a" Dec 02 13:23:14 crc kubenswrapper[4725]: E1202 13:23:14.947658 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a\": container with ID starting with 7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a not found: ID does not exist" containerID="7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.947698 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a"} err="failed to get container status \"7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a\": rpc error: code = NotFound desc = could not find container \"7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a\": container with ID starting with 7c87403997e82ce329763c4d8fb307d1da0857cee66028078b2db7cda9b6732a not found: ID does not exist" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.947713 4725 scope.go:117] "RemoveContainer" containerID="2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc" Dec 02 13:23:14 crc kubenswrapper[4725]: E1202 13:23:14.948959 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc\": container with ID starting with 2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc not found: ID does not exist" containerID="2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.948984 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc"} err="failed to get container status \"2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc\": rpc error: code = NotFound desc = could not find container \"2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc\": container with ID starting with 2391c060a0cf50b7d0256d9ced8619b87c09886a8efa7ad0b2858f1629d169bc not found: ID does not exist" Dec 02 13:23:14 crc kubenswrapper[4725]: I1202 13:23:14.948998 4725 scope.go:117] "RemoveContainer" containerID="46b16cc4c7d0e763786e97f65653f835629b8e641b67b0a6af1cb404505463d7" Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.248953 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b5b59474f-qnsqr" Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.326843 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="335fa14a-e2ec-4f9e-a56d-ab2a9be708f5" path="/var/lib/kubelet/pods/335fa14a-e2ec-4f9e-a56d-ab2a9be708f5/volumes" Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.330896 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc552b8a-fee3-4999-a174-c6505a51a9fd" path="/var/lib/kubelet/pods/cc552b8a-fee3-4999-a174-c6505a51a9fd/volumes" Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.331570 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64f9c49f9b-5bjhh"] Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.331942 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64f9c49f9b-5bjhh" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-api" containerID="cri-o://4ab58d5470d7602b49cc92f4e2bcbf04039ed1a49a679485db252829c8ab31cf" gracePeriod=30 Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.332097 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64f9c49f9b-5bjhh" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-httpd" containerID="cri-o://aa9195c8c0f9f6caee04b69871fa844b4d1778256decd817af290cf37baa99ed" gracePeriod=30 Dec 02 13:23:15 crc kubenswrapper[4725]: I1202 13:23:15.506593 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:15 crc kubenswrapper[4725]: W1202 13:23:15.508178 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod974e23a0_ac9d_419d_825e_b707b3808571.slice/crio-427e59d400265436baf265e774250bb5d4ae43e81f0f87bdd21d0b3cca9f7386 WatchSource:0}: Error finding container 427e59d400265436baf265e774250bb5d4ae43e81f0f87bdd21d0b3cca9f7386: Status 404 returned error can't find the container with id 427e59d400265436baf265e774250bb5d4ae43e81f0f87bdd21d0b3cca9f7386 Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.176546 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c6d675f88-kk42m"] Dec 02 13:23:16 crc kubenswrapper[4725]: E1202 13:23:16.177996 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc552b8a-fee3-4999-a174-c6505a51a9fd" containerName="init" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.178021 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc552b8a-fee3-4999-a174-c6505a51a9fd" containerName="init" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.178262 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc552b8a-fee3-4999-a174-c6505a51a9fd" containerName="init" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.179575 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.182969 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.183287 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.185790 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c6d675f88-kk42m"] Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.285492 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-config-data\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.287021 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-public-tls-certs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.287121 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-config-data-custom\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.287265 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8063dc7f-156e-4f2f-a6fa-8759abb28c37-logs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.287724 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-internal-tls-certs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.287908 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-combined-ca-bundle\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.288284 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg5sj\" (UniqueName: \"kubernetes.io/projected/8063dc7f-156e-4f2f-a6fa-8759abb28c37-kube-api-access-xg5sj\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.382203 4725 generic.go:334] "Generic (PLEG): container finished" podID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerID="aa9195c8c0f9f6caee04b69871fa844b4d1778256decd817af290cf37baa99ed" exitCode=0 Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.382358 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f9c49f9b-5bjhh" event={"ID":"66e1f474-e66c-4085-b7de-4f6ef80bfa84","Type":"ContainerDied","Data":"aa9195c8c0f9f6caee04b69871fa844b4d1778256decd817af290cf37baa99ed"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.389831 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-internal-tls-certs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.389880 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-combined-ca-bundle\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.389969 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg5sj\" (UniqueName: \"kubernetes.io/projected/8063dc7f-156e-4f2f-a6fa-8759abb28c37-kube-api-access-xg5sj\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.389998 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-config-data\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.390034 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-public-tls-certs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.390059 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-config-data-custom\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.390087 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8063dc7f-156e-4f2f-a6fa-8759abb28c37-logs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.391016 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fe1f66d-914d-4e77-9d98-e55681fdc408","Type":"ContainerStarted","Data":"d13d065111f5455751a240483eb7353878ef9eb549af4f40109caed469fb2793"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.391082 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8063dc7f-156e-4f2f-a6fa-8759abb28c37-logs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.399251 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-config-data-custom\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.399482 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-combined-ca-bundle\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.399711 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-config-data\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.399935 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-internal-tls-certs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.403144 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8063dc7f-156e-4f2f-a6fa-8759abb28c37-public-tls-certs\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.410397 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" event={"ID":"643b3918-6dab-4956-bd36-474ba0195d15","Type":"ContainerStarted","Data":"cece64e18f0df52274066ee2091aba55348e317b449c16f085dffffed7b861e1"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.410486 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" event={"ID":"643b3918-6dab-4956-bd36-474ba0195d15","Type":"ContainerStarted","Data":"f60755e3776c37dbf1e90bcd76d0c317692a47b574a1aadfb7409ff0bae452c4"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.415096 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg5sj\" (UniqueName: \"kubernetes.io/projected/8063dc7f-156e-4f2f-a6fa-8759abb28c37-kube-api-access-xg5sj\") pod \"barbican-api-6c6d675f88-kk42m\" (UID: \"8063dc7f-156e-4f2f-a6fa-8759abb28c37\") " pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.420814 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"16749490-6ff5-4ffc-bd23-d47ec3082d94","Type":"ContainerStarted","Data":"6b78954669c4301ae43f4b3a67ccd8528da7b93e04dff0d12fccadc3b98cef82"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.420904 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api-log" containerID="cri-o://c984a821133bc14a4ba4cad372309cb11a7e2488de1b4d8046e42c452c9ab432" gracePeriod=30 Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.420949 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.420993 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api" containerID="cri-o://6b78954669c4301ae43f4b3a67ccd8528da7b93e04dff0d12fccadc3b98cef82" gracePeriod=30 Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.432637 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5c5f8cd4f5-fd8q7" podStartSLOduration=3.890977026 podStartE2EDuration="7.43261785s" podCreationTimestamp="2025-12-02 13:23:09 +0000 UTC" firstStartedPulling="2025-12-02 13:23:11.46287617 +0000 UTC m=+1122.419517865" lastFinishedPulling="2025-12-02 13:23:15.004516994 +0000 UTC m=+1125.961158689" observedRunningTime="2025-12-02 13:23:16.430951328 +0000 UTC m=+1127.387593023" watchObservedRunningTime="2025-12-02 13:23:16.43261785 +0000 UTC m=+1127.389259545" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.485428 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.48539607 podStartE2EDuration="6.48539607s" podCreationTimestamp="2025-12-02 13:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:16.470857156 +0000 UTC m=+1127.427498851" watchObservedRunningTime="2025-12-02 13:23:16.48539607 +0000 UTC m=+1127.442037765" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.501641 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerStarted","Data":"427e59d400265436baf265e774250bb5d4ae43e81f0f87bdd21d0b3cca9f7386"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.531841 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.534833 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" event={"ID":"0e621595-2643-4b38-913c-23bc8f80fb62","Type":"ContainerStarted","Data":"5bc681abf46687702379f1e084a10192f803233e7ce723b8f9138dde893032de"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.535732 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.546742 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" event={"ID":"e5f2ea09-3acf-4c71-8301-c6531e608bf1","Type":"ContainerStarted","Data":"f99e203d7ad143c703f8b8ee9c413e5aff7844d8d66f0ef32a42de185ff63b99"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.546802 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" event={"ID":"e5f2ea09-3acf-4c71-8301-c6531e608bf1","Type":"ContainerStarted","Data":"31fb3fade4f87afc8200f6229712b6454c0f704e7c728cf43fb6bac8b0747885"} Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.566836 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" podStartSLOduration=6.566809398 podStartE2EDuration="6.566809398s" podCreationTimestamp="2025-12-02 13:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:16.561430533 +0000 UTC m=+1127.518072248" watchObservedRunningTime="2025-12-02 13:23:16.566809398 +0000 UTC m=+1127.523451093" Dec 02 13:23:16 crc kubenswrapper[4725]: I1202 13:23:16.614733 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-54b4d7cbbb-hdq62" podStartSLOduration=3.403186112 podStartE2EDuration="6.614625154s" podCreationTimestamp="2025-12-02 13:23:10 +0000 UTC" firstStartedPulling="2025-12-02 13:23:11.818804826 +0000 UTC m=+1122.775446521" lastFinishedPulling="2025-12-02 13:23:15.030243868 +0000 UTC m=+1125.986885563" observedRunningTime="2025-12-02 13:23:16.587506945 +0000 UTC m=+1127.544148640" watchObservedRunningTime="2025-12-02 13:23:16.614625154 +0000 UTC m=+1127.571266879" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.417996 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c6d675f88-kk42m"] Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.566578 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c6d675f88-kk42m" event={"ID":"8063dc7f-156e-4f2f-a6fa-8759abb28c37","Type":"ContainerStarted","Data":"a6f94e974c161b87e95ac49f32b296b75a2a032610d8db57acab11e1a78f8cab"} Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.568264 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerStarted","Data":"d4729d1ae15d6e2d3424c9aee5febe575496578a32db695a550e2e92d73102a1"} Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.571535 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fe1f66d-914d-4e77-9d98-e55681fdc408","Type":"ContainerStarted","Data":"09d14b9a1614f272d2368b47fa1d8d3ad57fff99a85a44c0c5583ac9e8bcf3d2"} Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.594966 4725 generic.go:334] "Generic (PLEG): container finished" podID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerID="6b78954669c4301ae43f4b3a67ccd8528da7b93e04dff0d12fccadc3b98cef82" exitCode=0 Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.595032 4725 generic.go:334] "Generic (PLEG): container finished" podID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerID="c984a821133bc14a4ba4cad372309cb11a7e2488de1b4d8046e42c452c9ab432" exitCode=143 Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.595490 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"16749490-6ff5-4ffc-bd23-d47ec3082d94","Type":"ContainerDied","Data":"6b78954669c4301ae43f4b3a67ccd8528da7b93e04dff0d12fccadc3b98cef82"} Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.595601 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"16749490-6ff5-4ffc-bd23-d47ec3082d94","Type":"ContainerDied","Data":"c984a821133bc14a4ba4cad372309cb11a7e2488de1b4d8046e42c452c9ab432"} Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.600963 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.251760617 podStartE2EDuration="8.600942266s" podCreationTimestamp="2025-12-02 13:23:09 +0000 UTC" firstStartedPulling="2025-12-02 13:23:11.819087103 +0000 UTC m=+1122.775728798" lastFinishedPulling="2025-12-02 13:23:15.168268742 +0000 UTC m=+1126.124910447" observedRunningTime="2025-12-02 13:23:17.598845753 +0000 UTC m=+1128.555487448" watchObservedRunningTime="2025-12-02 13:23:17.600942266 +0000 UTC m=+1128.557583961" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.806000 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828305 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-scripts\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828384 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828522 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-combined-ca-bundle\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828583 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msxlv\" (UniqueName: \"kubernetes.io/projected/16749490-6ff5-4ffc-bd23-d47ec3082d94-kube-api-access-msxlv\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828638 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data-custom\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828703 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16749490-6ff5-4ffc-bd23-d47ec3082d94-etc-machine-id\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.828755 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16749490-6ff5-4ffc-bd23-d47ec3082d94-logs\") pod \"16749490-6ff5-4ffc-bd23-d47ec3082d94\" (UID: \"16749490-6ff5-4ffc-bd23-d47ec3082d94\") " Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.829061 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/16749490-6ff5-4ffc-bd23-d47ec3082d94-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.829204 4725 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/16749490-6ff5-4ffc-bd23-d47ec3082d94-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.829667 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16749490-6ff5-4ffc-bd23-d47ec3082d94-logs" (OuterVolumeSpecName: "logs") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.840076 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-scripts" (OuterVolumeSpecName: "scripts") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.845910 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16749490-6ff5-4ffc-bd23-d47ec3082d94-kube-api-access-msxlv" (OuterVolumeSpecName: "kube-api-access-msxlv") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "kube-api-access-msxlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.861732 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.919833 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.931276 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msxlv\" (UniqueName: \"kubernetes.io/projected/16749490-6ff5-4ffc-bd23-d47ec3082d94-kube-api-access-msxlv\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.931347 4725 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.931359 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16749490-6ff5-4ffc-bd23-d47ec3082d94-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.931370 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.931380 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:17 crc kubenswrapper[4725]: I1202 13:23:17.976293 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data" (OuterVolumeSpecName: "config-data") pod "16749490-6ff5-4ffc-bd23-d47ec3082d94" (UID: "16749490-6ff5-4ffc-bd23-d47ec3082d94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.034038 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16749490-6ff5-4ffc-bd23-d47ec3082d94-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.614414 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.615484 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"16749490-6ff5-4ffc-bd23-d47ec3082d94","Type":"ContainerDied","Data":"118f676d2d63f8f5bf1539e85d35ca63a2970ea018f6a797ed4c82e506daf731"} Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.615994 4725 scope.go:117] "RemoveContainer" containerID="6b78954669c4301ae43f4b3a67ccd8528da7b93e04dff0d12fccadc3b98cef82" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.638727 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c6d675f88-kk42m" event={"ID":"8063dc7f-156e-4f2f-a6fa-8759abb28c37","Type":"ContainerStarted","Data":"b83f20286eb11e82caceb2129997e855e5ed58f56387a398e138a5ebc26104c2"} Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.638776 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c6d675f88-kk42m" event={"ID":"8063dc7f-156e-4f2f-a6fa-8759abb28c37","Type":"ContainerStarted","Data":"2727af372ceccd7c256c0afb334d7d3b572b2f02c1e2aacbe3b3def439fcf579"} Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.639476 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.639680 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.645900 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerStarted","Data":"e6a7f1038a30b5c4e15914bb4cf0a28df74bb6479e0b93e7663f26a6ebb615c2"} Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.665594 4725 scope.go:117] "RemoveContainer" containerID="c984a821133bc14a4ba4cad372309cb11a7e2488de1b4d8046e42c452c9ab432" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.673583 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c6d675f88-kk42m" podStartSLOduration=2.673551367 podStartE2EDuration="2.673551367s" podCreationTimestamp="2025-12-02 13:23:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:18.661727011 +0000 UTC m=+1129.618368706" watchObservedRunningTime="2025-12-02 13:23:18.673551367 +0000 UTC m=+1129.630193062" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.699522 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.747003 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.798053 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:18 crc kubenswrapper[4725]: E1202 13:23:18.798588 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.798606 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api" Dec 02 13:23:18 crc kubenswrapper[4725]: E1202 13:23:18.798623 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api-log" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.798630 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api-log" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.798841 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.798854 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" containerName="cinder-api-log" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.800005 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.810778 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.811808 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.812015 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.844915 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865079 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-config-data\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865182 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eda48a5c-28d1-4b25-8377-a8870261551c-logs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865283 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865327 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eda48a5c-28d1-4b25-8377-a8870261551c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865375 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865478 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865606 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h9b2\" (UniqueName: \"kubernetes.io/projected/eda48a5c-28d1-4b25-8377-a8870261551c-kube-api-access-2h9b2\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865645 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-config-data-custom\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.865687 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-scripts\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.967710 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eda48a5c-28d1-4b25-8377-a8870261551c-logs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.967807 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.967848 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eda48a5c-28d1-4b25-8377-a8870261551c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.967882 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.967923 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.967997 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h9b2\" (UniqueName: \"kubernetes.io/projected/eda48a5c-28d1-4b25-8377-a8870261551c-kube-api-access-2h9b2\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.968025 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eda48a5c-28d1-4b25-8377-a8870261551c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.968034 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-config-data-custom\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.968174 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-scripts\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.968257 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eda48a5c-28d1-4b25-8377-a8870261551c-logs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.968269 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-config-data\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.973424 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.974053 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.976849 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-scripts\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:18 crc kubenswrapper[4725]: I1202 13:23:18.987165 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-config-data-custom\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.005348 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-config-data\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.006012 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eda48a5c-28d1-4b25-8377-a8870261551c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.006868 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h9b2\" (UniqueName: \"kubernetes.io/projected/eda48a5c-28d1-4b25-8377-a8870261551c-kube-api-access-2h9b2\") pod \"cinder-api-0\" (UID: \"eda48a5c-28d1-4b25-8377-a8870261551c\") " pod="openstack/cinder-api-0" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.137578 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.294617 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16749490-6ff5-4ffc-bd23-d47ec3082d94" path="/var/lib/kubelet/pods/16749490-6ff5-4ffc-bd23-d47ec3082d94/volumes" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.666332 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerStarted","Data":"8621851f97560df512e8694c05b8af2faa32a78deb1c32b54ad0ab978dc58a44"} Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.667027 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.673250 4725 generic.go:334] "Generic (PLEG): container finished" podID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerID="4ab58d5470d7602b49cc92f4e2bcbf04039ed1a49a679485db252829c8ab31cf" exitCode=0 Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.674640 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f9c49f9b-5bjhh" event={"ID":"66e1f474-e66c-4085-b7de-4f6ef80bfa84","Type":"ContainerDied","Data":"4ab58d5470d7602b49cc92f4e2bcbf04039ed1a49a679485db252829c8ab31cf"} Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.674684 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64f9c49f9b-5bjhh" event={"ID":"66e1f474-e66c-4085-b7de-4f6ef80bfa84","Type":"ContainerDied","Data":"7ea24c92f55e14f352638872ef5ce8eab1e1aea12abe60db0aa7624e9e502906"} Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.674702 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ea24c92f55e14f352638872ef5ce8eab1e1aea12abe60db0aa7624e9e502906" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.718552 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.796176 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-config\") pod \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.796265 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8jvx\" (UniqueName: \"kubernetes.io/projected/66e1f474-e66c-4085-b7de-4f6ef80bfa84-kube-api-access-t8jvx\") pod \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.796296 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-httpd-config\") pod \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.805341 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e1f474-e66c-4085-b7de-4f6ef80bfa84-kube-api-access-t8jvx" (OuterVolumeSpecName: "kube-api-access-t8jvx") pod "66e1f474-e66c-4085-b7de-4f6ef80bfa84" (UID: "66e1f474-e66c-4085-b7de-4f6ef80bfa84"). InnerVolumeSpecName "kube-api-access-t8jvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.807021 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "66e1f474-e66c-4085-b7de-4f6ef80bfa84" (UID: "66e1f474-e66c-4085-b7de-4f6ef80bfa84"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.829247 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.891276 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-config" (OuterVolumeSpecName: "config") pod "66e1f474-e66c-4085-b7de-4f6ef80bfa84" (UID: "66e1f474-e66c-4085-b7de-4f6ef80bfa84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.900485 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-combined-ca-bundle\") pod \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.901290 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-ovndb-tls-certs\") pod \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\" (UID: \"66e1f474-e66c-4085-b7de-4f6ef80bfa84\") " Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.901960 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.901988 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8jvx\" (UniqueName: \"kubernetes.io/projected/66e1f474-e66c-4085-b7de-4f6ef80bfa84-kube-api-access-t8jvx\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:19 crc kubenswrapper[4725]: I1202 13:23:19.902007 4725 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.010796 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66e1f474-e66c-4085-b7de-4f6ef80bfa84" (UID: "66e1f474-e66c-4085-b7de-4f6ef80bfa84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.014585 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.131002 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "66e1f474-e66c-4085-b7de-4f6ef80bfa84" (UID: "66e1f474-e66c-4085-b7de-4f6ef80bfa84"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.227869 4725 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/66e1f474-e66c-4085-b7de-4f6ef80bfa84-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.390296 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.437758 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.693980 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eda48a5c-28d1-4b25-8377-a8870261551c","Type":"ContainerStarted","Data":"fa4abe66f7a351cee936ee5551732fab3f1ca30700ad677d1283ec37e0570092"} Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.694044 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64f9c49f9b-5bjhh" Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.743640 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64f9c49f9b-5bjhh"] Dec 02 13:23:20 crc kubenswrapper[4725]: I1202 13:23:20.754770 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-64f9c49f9b-5bjhh"] Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.036700 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.141305 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vmglp"] Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.142971 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="dnsmasq-dns" containerID="cri-o://40da217e5fc67e9c37dfc1b392f5ffeaa1136f22348a1786f35addb428b40375" gracePeriod=10 Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.349483 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" path="/var/lib/kubelet/pods/66e1f474-e66c-4085-b7de-4f6ef80bfa84/volumes" Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.731628 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eda48a5c-28d1-4b25-8377-a8870261551c","Type":"ContainerStarted","Data":"6104054a4fa19e0adcabc4e348df52ffcdb8844bd5daf97ff967a5b32f128309"} Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.741219 4725 generic.go:334] "Generic (PLEG): container finished" podID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerID="40da217e5fc67e9c37dfc1b392f5ffeaa1136f22348a1786f35addb428b40375" exitCode=0 Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.741601 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" event={"ID":"089c681d-b6d5-43ac-9fca-ed2a92d59441","Type":"ContainerDied","Data":"40da217e5fc67e9c37dfc1b392f5ffeaa1136f22348a1786f35addb428b40375"} Dec 02 13:23:21 crc kubenswrapper[4725]: I1202 13:23:21.987491 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.076988 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z5zz\" (UniqueName: \"kubernetes.io/projected/089c681d-b6d5-43ac-9fca-ed2a92d59441-kube-api-access-5z5zz\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.077196 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-nb\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.077226 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-svc\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.077276 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.077329 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-config\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.077368 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-sb\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.113937 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089c681d-b6d5-43ac-9fca-ed2a92d59441-kube-api-access-5z5zz" (OuterVolumeSpecName: "kube-api-access-5z5zz") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "kube-api-access-5z5zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.179824 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z5zz\" (UniqueName: \"kubernetes.io/projected/089c681d-b6d5-43ac-9fca-ed2a92d59441-kube-api-access-5z5zz\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.370936 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.378339 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f9f547644-22mp9" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.381709 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-config" (OuterVolumeSpecName: "config") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.382452 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.383337 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0\") pod \"089c681d-b6d5-43ac-9fca-ed2a92d59441\" (UID: \"089c681d-b6d5-43ac-9fca-ed2a92d59441\") " Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.383838 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.383857 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.383887 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.385704 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: W1202 13:23:22.386182 4725 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/089c681d-b6d5-43ac-9fca-ed2a92d59441/volumes/kubernetes.io~configmap/dns-swift-storage-0 Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.386381 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "089c681d-b6d5-43ac-9fca-ed2a92d59441" (UID: "089c681d-b6d5-43ac-9fca-ed2a92d59441"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.484281 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f44bd87bd-2w6jw"] Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.484537 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon-log" containerID="cri-o://149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077" gracePeriod=30 Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.485244 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" containerID="cri-o://0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360" gracePeriod=30 Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.485649 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.485674 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.485683 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/089c681d-b6d5-43ac-9fca-ed2a92d59441-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.495707 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.725271 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.794640 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerStarted","Data":"905449c49700f8ba5cf3d89838967027a8981633830d93c4b1abf9ae14044138"} Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.798341 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" event={"ID":"089c681d-b6d5-43ac-9fca-ed2a92d59441","Type":"ContainerDied","Data":"04500d9780fc70441529ef6d545d41a54f52decbf79da38b5a08c0b19f8959aa"} Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.798554 4725 scope.go:117] "RemoveContainer" containerID="40da217e5fc67e9c37dfc1b392f5ffeaa1136f22348a1786f35addb428b40375" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.798431 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.810556 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"eda48a5c-28d1-4b25-8377-a8870261551c","Type":"ContainerStarted","Data":"fa5b73a4e6aeab8c75f889df88ff86b12e5730091ed1c864138c06d231eb7b19"} Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.837988 4725 scope.go:117] "RemoveContainer" containerID="5c8d5fff6a51feb41575531bc7f2c2030eaa343034944857c8509c81bdc2bdb2" Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.865610 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vmglp"] Dec 02 13:23:22 crc kubenswrapper[4725]: I1202 13:23:22.871317 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vmglp"] Dec 02 13:23:23 crc kubenswrapper[4725]: I1202 13:23:23.236958 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:23 crc kubenswrapper[4725]: I1202 13:23:23.278873 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" path="/var/lib/kubelet/pods/089c681d-b6d5-43ac-9fca-ed2a92d59441/volumes" Dec 02 13:23:23 crc kubenswrapper[4725]: I1202 13:23:23.824642 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 02 13:23:23 crc kubenswrapper[4725]: I1202 13:23:23.824938 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:23:23 crc kubenswrapper[4725]: I1202 13:23:23.849474 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.700708672 podStartE2EDuration="10.849437366s" podCreationTimestamp="2025-12-02 13:23:13 +0000 UTC" firstStartedPulling="2025-12-02 13:23:15.512046404 +0000 UTC m=+1126.468688099" lastFinishedPulling="2025-12-02 13:23:21.660775098 +0000 UTC m=+1132.617416793" observedRunningTime="2025-12-02 13:23:23.845749524 +0000 UTC m=+1134.802391249" watchObservedRunningTime="2025-12-02 13:23:23.849437366 +0000 UTC m=+1134.806079061" Dec 02 13:23:23 crc kubenswrapper[4725]: I1202 13:23:23.876543 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.876524354 podStartE2EDuration="5.876524354s" podCreationTimestamp="2025-12-02 13:23:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:23.874132254 +0000 UTC m=+1134.830773959" watchObservedRunningTime="2025-12-02 13:23:23.876524354 +0000 UTC m=+1134.833166049" Dec 02 13:23:24 crc kubenswrapper[4725]: I1202 13:23:24.284962 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:23:24 crc kubenswrapper[4725]: I1202 13:23:24.285020 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:23:25 crc kubenswrapper[4725]: I1202 13:23:25.593187 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c7dc85fc9-tj9wl" Dec 02 13:23:25 crc kubenswrapper[4725]: I1202 13:23:25.719281 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 02 13:23:25 crc kubenswrapper[4725]: I1202 13:23:25.768589 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:25 crc kubenswrapper[4725]: I1202 13:23:25.852784 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="cinder-scheduler" containerID="cri-o://d13d065111f5455751a240483eb7353878ef9eb549af4f40109caed469fb2793" gracePeriod=30 Dec 02 13:23:25 crc kubenswrapper[4725]: I1202 13:23:25.852991 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="probe" containerID="cri-o://09d14b9a1614f272d2368b47fa1d8d3ad57fff99a85a44c0c5583ac9e8bcf3d2" gracePeriod=30 Dec 02 13:23:25 crc kubenswrapper[4725]: I1202 13:23:25.916342 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:51104->10.217.0.147:8443: read: connection reset by peer" Dec 02 13:23:26 crc kubenswrapper[4725]: I1202 13:23:26.621352 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 02 13:23:26 crc kubenswrapper[4725]: I1202 13:23:26.894118 4725 generic.go:334] "Generic (PLEG): container finished" podID="cd102f85-1b92-4215-95e9-33baf5949dec" containerID="0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360" exitCode=0 Dec 02 13:23:26 crc kubenswrapper[4725]: I1202 13:23:26.894205 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44bd87bd-2w6jw" event={"ID":"cd102f85-1b92-4215-95e9-33baf5949dec","Type":"ContainerDied","Data":"0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360"} Dec 02 13:23:26 crc kubenswrapper[4725]: I1202 13:23:26.951797 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-vmglp" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: i/o timeout" Dec 02 13:23:27 crc kubenswrapper[4725]: I1202 13:23:27.929261 4725 generic.go:334] "Generic (PLEG): container finished" podID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerID="09d14b9a1614f272d2368b47fa1d8d3ad57fff99a85a44c0c5583ac9e8bcf3d2" exitCode=0 Dec 02 13:23:27 crc kubenswrapper[4725]: I1202 13:23:27.929334 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fe1f66d-914d-4e77-9d98-e55681fdc408","Type":"ContainerDied","Data":"09d14b9a1614f272d2368b47fa1d8d3ad57fff99a85a44c0c5583ac9e8bcf3d2"} Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.581122 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 02 13:23:28 crc kubenswrapper[4725]: E1202 13:23:28.582200 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-api" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582224 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-api" Dec 02 13:23:28 crc kubenswrapper[4725]: E1202 13:23:28.582240 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="dnsmasq-dns" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582247 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="dnsmasq-dns" Dec 02 13:23:28 crc kubenswrapper[4725]: E1202 13:23:28.582275 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-httpd" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582281 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-httpd" Dec 02 13:23:28 crc kubenswrapper[4725]: E1202 13:23:28.582316 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="init" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582322 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="init" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582520 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="089c681d-b6d5-43ac-9fca-ed2a92d59441" containerName="dnsmasq-dns" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582539 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-httpd" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.582565 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e1f474-e66c-4085-b7de-4f6ef80bfa84" containerName="neutron-api" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.583358 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.591809 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-7tvn7" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.591922 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.591923 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.615378 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.756856 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjqbh\" (UniqueName: \"kubernetes.io/projected/6ed64991-e56a-4635-b062-f68f7f1d134d-kube-api-access-bjqbh\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.756945 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6ed64991-e56a-4635-b062-f68f7f1d134d-openstack-config-secret\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.756989 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6ed64991-e56a-4635-b062-f68f7f1d134d-openstack-config\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.757067 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed64991-e56a-4635-b062-f68f7f1d134d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.859367 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjqbh\" (UniqueName: \"kubernetes.io/projected/6ed64991-e56a-4635-b062-f68f7f1d134d-kube-api-access-bjqbh\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.859476 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6ed64991-e56a-4635-b062-f68f7f1d134d-openstack-config-secret\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.859511 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6ed64991-e56a-4635-b062-f68f7f1d134d-openstack-config\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.859577 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed64991-e56a-4635-b062-f68f7f1d134d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.860990 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6ed64991-e56a-4635-b062-f68f7f1d134d-openstack-config\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.877942 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ed64991-e56a-4635-b062-f68f7f1d134d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.881670 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6ed64991-e56a-4635-b062-f68f7f1d134d-openstack-config-secret\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.885990 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjqbh\" (UniqueName: \"kubernetes.io/projected/6ed64991-e56a-4635-b062-f68f7f1d134d-kube-api-access-bjqbh\") pod \"openstackclient\" (UID: \"6ed64991-e56a-4635-b062-f68f7f1d134d\") " pod="openstack/openstackclient" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.904723 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:28 crc kubenswrapper[4725]: I1202 13:23:28.915271 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.461516 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.581246 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c6d675f88-kk42m" Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.659331 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6644cd5fd-44gdx"] Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.659973 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6644cd5fd-44gdx" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api-log" containerID="cri-o://ac34e06a59152d1275189c8bd1313e1f87feffac653b3ac6a7ccb2f3aac0a747" gracePeriod=30 Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.660386 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6644cd5fd-44gdx" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api" containerID="cri-o://05e658ba9914f6b64a7afc3562df07ebc1c68f21d618f6a545cac5c653e83aba" gracePeriod=30 Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.977914 4725 generic.go:334] "Generic (PLEG): container finished" podID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerID="ac34e06a59152d1275189c8bd1313e1f87feffac653b3ac6a7ccb2f3aac0a747" exitCode=143 Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.978033 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6644cd5fd-44gdx" event={"ID":"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948","Type":"ContainerDied","Data":"ac34e06a59152d1275189c8bd1313e1f87feffac653b3ac6a7ccb2f3aac0a747"} Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.987428 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6ed64991-e56a-4635-b062-f68f7f1d134d","Type":"ContainerStarted","Data":"d7a1cbcc35e2d1603a1d4ddb2ad938e214a7ed2e98f7ee50e24119f9deb5c8df"} Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.994380 4725 generic.go:334] "Generic (PLEG): container finished" podID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerID="d13d065111f5455751a240483eb7353878ef9eb549af4f40109caed469fb2793" exitCode=0 Dec 02 13:23:29 crc kubenswrapper[4725]: I1202 13:23:29.994443 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fe1f66d-914d-4e77-9d98-e55681fdc408","Type":"ContainerDied","Data":"d13d065111f5455751a240483eb7353878ef9eb549af4f40109caed469fb2793"} Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.248575 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.387563 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-scripts\") pod \"1fe1f66d-914d-4e77-9d98-e55681fdc408\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.387655 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwghx\" (UniqueName: \"kubernetes.io/projected/1fe1f66d-914d-4e77-9d98-e55681fdc408-kube-api-access-gwghx\") pod \"1fe1f66d-914d-4e77-9d98-e55681fdc408\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.387733 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data-custom\") pod \"1fe1f66d-914d-4e77-9d98-e55681fdc408\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.387808 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data\") pod \"1fe1f66d-914d-4e77-9d98-e55681fdc408\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.387944 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fe1f66d-914d-4e77-9d98-e55681fdc408-etc-machine-id\") pod \"1fe1f66d-914d-4e77-9d98-e55681fdc408\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.388054 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-combined-ca-bundle\") pod \"1fe1f66d-914d-4e77-9d98-e55681fdc408\" (UID: \"1fe1f66d-914d-4e77-9d98-e55681fdc408\") " Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.389273 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1fe1f66d-914d-4e77-9d98-e55681fdc408-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1fe1f66d-914d-4e77-9d98-e55681fdc408" (UID: "1fe1f66d-914d-4e77-9d98-e55681fdc408"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.390036 4725 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fe1f66d-914d-4e77-9d98-e55681fdc408-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.406263 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1fe1f66d-914d-4e77-9d98-e55681fdc408" (UID: "1fe1f66d-914d-4e77-9d98-e55681fdc408"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.409034 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-scripts" (OuterVolumeSpecName: "scripts") pod "1fe1f66d-914d-4e77-9d98-e55681fdc408" (UID: "1fe1f66d-914d-4e77-9d98-e55681fdc408"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.431843 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fe1f66d-914d-4e77-9d98-e55681fdc408-kube-api-access-gwghx" (OuterVolumeSpecName: "kube-api-access-gwghx") pod "1fe1f66d-914d-4e77-9d98-e55681fdc408" (UID: "1fe1f66d-914d-4e77-9d98-e55681fdc408"). InnerVolumeSpecName "kube-api-access-gwghx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.497124 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.497160 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwghx\" (UniqueName: \"kubernetes.io/projected/1fe1f66d-914d-4e77-9d98-e55681fdc408-kube-api-access-gwghx\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.497191 4725 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.536239 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fe1f66d-914d-4e77-9d98-e55681fdc408" (UID: "1fe1f66d-914d-4e77-9d98-e55681fdc408"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.585064 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data" (OuterVolumeSpecName: "config-data") pod "1fe1f66d-914d-4e77-9d98-e55681fdc408" (UID: "1fe1f66d-914d-4e77-9d98-e55681fdc408"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.599560 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:30 crc kubenswrapper[4725]: I1202 13:23:30.599616 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fe1f66d-914d-4e77-9d98-e55681fdc408-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.017435 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fe1f66d-914d-4e77-9d98-e55681fdc408","Type":"ContainerDied","Data":"f95c3f67c3f6971d44d736a9b67e0538471772a3e021193d6c92195984d435fe"} Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.017577 4725 scope.go:117] "RemoveContainer" containerID="09d14b9a1614f272d2368b47fa1d8d3ad57fff99a85a44c0c5583ac9e8bcf3d2" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.017514 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.068581 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.083125 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.101385 4725 scope.go:117] "RemoveContainer" containerID="d13d065111f5455751a240483eb7353878ef9eb549af4f40109caed469fb2793" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.109642 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:31 crc kubenswrapper[4725]: E1202 13:23:31.110098 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="probe" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.110111 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="probe" Dec 02 13:23:31 crc kubenswrapper[4725]: E1202 13:23:31.110135 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="cinder-scheduler" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.110142 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="cinder-scheduler" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.110332 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="cinder-scheduler" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.110342 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" containerName="probe" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.111312 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.123804 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.123931 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-scripts\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.124000 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxzp5\" (UniqueName: \"kubernetes.io/projected/3eab134b-db07-44ea-9f46-734dbda09036-kube-api-access-bxzp5\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.124035 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.124071 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3eab134b-db07-44ea-9f46-734dbda09036-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.124111 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-config-data\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.124231 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.147624 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.226739 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.226859 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-scripts\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.226932 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxzp5\" (UniqueName: \"kubernetes.io/projected/3eab134b-db07-44ea-9f46-734dbda09036-kube-api-access-bxzp5\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.226973 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.227017 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3eab134b-db07-44ea-9f46-734dbda09036-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.227060 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-config-data\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.228391 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3eab134b-db07-44ea-9f46-734dbda09036-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.232859 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.235643 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-config-data\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.236323 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.242326 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eab134b-db07-44ea-9f46-734dbda09036-scripts\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.253357 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxzp5\" (UniqueName: \"kubernetes.io/projected/3eab134b-db07-44ea-9f46-734dbda09036-kube-api-access-bxzp5\") pod \"cinder-scheduler-0\" (UID: \"3eab134b-db07-44ea-9f46-734dbda09036\") " pod="openstack/cinder-scheduler-0" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.290194 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fe1f66d-914d-4e77-9d98-e55681fdc408" path="/var/lib/kubelet/pods/1fe1f66d-914d-4e77-9d98-e55681fdc408/volumes" Dec 02 13:23:31 crc kubenswrapper[4725]: I1202 13:23:31.452017 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 13:23:32 crc kubenswrapper[4725]: I1202 13:23:32.090914 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 13:23:33 crc kubenswrapper[4725]: I1202 13:23:33.051122 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3eab134b-db07-44ea-9f46-734dbda09036","Type":"ContainerStarted","Data":"998cdb874861b5b1b2caafe60f4b22813dd3d0bc634f9ff5267c5228f51aa72d"} Dec 02 13:23:33 crc kubenswrapper[4725]: I1202 13:23:33.087339 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 02 13:23:33 crc kubenswrapper[4725]: I1202 13:23:33.701987 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6644cd5fd-44gdx" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:50162->10.217.0.161:9311: read: connection reset by peer" Dec 02 13:23:33 crc kubenswrapper[4725]: I1202 13:23:33.701999 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6644cd5fd-44gdx" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": read tcp 10.217.0.2:50170->10.217.0.161:9311: read: connection reset by peer" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.080625 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3eab134b-db07-44ea-9f46-734dbda09036","Type":"ContainerStarted","Data":"a82e6deace5e610c49de5de3997398c83fe6d97cbfe7d2c99e6acf91fe042559"} Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.093773 4725 generic.go:334] "Generic (PLEG): container finished" podID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerID="05e658ba9914f6b64a7afc3562df07ebc1c68f21d618f6a545cac5c653e83aba" exitCode=0 Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.093915 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6644cd5fd-44gdx" event={"ID":"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948","Type":"ContainerDied","Data":"05e658ba9914f6b64a7afc3562df07ebc1c68f21d618f6a545cac5c653e83aba"} Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.255893 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.423341 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6db82\" (UniqueName: \"kubernetes.io/projected/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-kube-api-access-6db82\") pod \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.423448 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data-custom\") pod \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.423493 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-combined-ca-bundle\") pod \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.423524 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data\") pod \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.423586 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-logs\") pod \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\" (UID: \"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948\") " Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.424813 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-logs" (OuterVolumeSpecName: "logs") pod "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" (UID: "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.472659 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" (UID: "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.476597 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-kube-api-access-6db82" (OuterVolumeSpecName: "kube-api-access-6db82") pod "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" (UID: "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948"). InnerVolumeSpecName "kube-api-access-6db82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.482301 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" (UID: "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.506993 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data" (OuterVolumeSpecName: "config-data") pod "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" (UID: "f0fb6745-6c04-4e3f-a59e-c6d85ebd1948"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.525863 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6db82\" (UniqueName: \"kubernetes.io/projected/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-kube-api-access-6db82\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.525897 4725 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.525907 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.525916 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:34 crc kubenswrapper[4725]: I1202 13:23:34.525927 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.109140 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3eab134b-db07-44ea-9f46-734dbda09036","Type":"ContainerStarted","Data":"a0d36c583b92c3a55e1880f34a9deef22afc4457baf1f8349fb75a18aaee0755"} Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.115572 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6644cd5fd-44gdx" event={"ID":"f0fb6745-6c04-4e3f-a59e-c6d85ebd1948","Type":"ContainerDied","Data":"6630ad5d73defc990f8549a21033b72ea9b721a8c012a582c40608896569d400"} Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.115616 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6644cd5fd-44gdx" Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.115669 4725 scope.go:117] "RemoveContainer" containerID="05e658ba9914f6b64a7afc3562df07ebc1c68f21d618f6a545cac5c653e83aba" Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.141184 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.141167777 podStartE2EDuration="4.141167777s" podCreationTimestamp="2025-12-02 13:23:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:35.136443179 +0000 UTC m=+1146.093084884" watchObservedRunningTime="2025-12-02 13:23:35.141167777 +0000 UTC m=+1146.097809472" Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.163213 4725 scope.go:117] "RemoveContainer" containerID="ac34e06a59152d1275189c8bd1313e1f87feffac653b3ac6a7ccb2f3aac0a747" Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.164690 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6644cd5fd-44gdx"] Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.184281 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6644cd5fd-44gdx"] Dec 02 13:23:35 crc kubenswrapper[4725]: I1202 13:23:35.290657 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" path="/var/lib/kubelet/pods/f0fb6745-6c04-4e3f-a59e-c6d85ebd1948/volumes" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.081546 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6d448f6767-ccmmh"] Dec 02 13:23:36 crc kubenswrapper[4725]: E1202 13:23:36.082650 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api-log" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.082680 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api-log" Dec 02 13:23:36 crc kubenswrapper[4725]: E1202 13:23:36.082713 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.082726 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.083002 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.083036 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0fb6745-6c04-4e3f-a59e-c6d85ebd1948" containerName="barbican-api-log" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.084729 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.091400 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.091564 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.092584 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.101812 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d448f6767-ccmmh"] Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.171950 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-config-data\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.172100 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvklk\" (UniqueName: \"kubernetes.io/projected/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-kube-api-access-rvklk\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.172141 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-combined-ca-bundle\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.172179 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-public-tls-certs\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.172247 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-run-httpd\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.174397 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-etc-swift\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.174519 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-internal-tls-certs\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.174569 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-log-httpd\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276317 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-etc-swift\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276439 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-internal-tls-certs\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276492 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-log-httpd\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276515 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-config-data\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276551 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvklk\" (UniqueName: \"kubernetes.io/projected/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-kube-api-access-rvklk\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276571 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-combined-ca-bundle\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276593 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-public-tls-certs\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.276635 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-run-httpd\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.278096 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-log-httpd\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.278188 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-run-httpd\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.288478 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-internal-tls-certs\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.289398 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-config-data\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.289682 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-etc-swift\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.293424 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-combined-ca-bundle\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.293445 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-public-tls-certs\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.297182 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvklk\" (UniqueName: \"kubernetes.io/projected/69be1bb2-b1e2-41fb-991c-c6a12db18fb8-kube-api-access-rvklk\") pod \"swift-proxy-6d448f6767-ccmmh\" (UID: \"69be1bb2-b1e2-41fb-991c-c6a12db18fb8\") " pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.453781 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.471946 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:36 crc kubenswrapper[4725]: I1202 13:23:36.619262 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 02 13:23:37 crc kubenswrapper[4725]: I1202 13:23:37.349515 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6d448f6767-ccmmh"] Dec 02 13:23:37 crc kubenswrapper[4725]: W1202 13:23:37.360635 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69be1bb2_b1e2_41fb_991c_c6a12db18fb8.slice/crio-8384913fd8523a85d2171dbb65482934a36328cfc074aa71f8e70c78d9068d00 WatchSource:0}: Error finding container 8384913fd8523a85d2171dbb65482934a36328cfc074aa71f8e70c78d9068d00: Status 404 returned error can't find the container with id 8384913fd8523a85d2171dbb65482934a36328cfc074aa71f8e70c78d9068d00 Dec 02 13:23:38 crc kubenswrapper[4725]: I1202 13:23:38.211111 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d448f6767-ccmmh" event={"ID":"69be1bb2-b1e2-41fb-991c-c6a12db18fb8","Type":"ContainerStarted","Data":"7f0e39f54d32e8c66cd414da7a0fbd32c0a4fd9747f305408434cc396395bad1"} Dec 02 13:23:38 crc kubenswrapper[4725]: I1202 13:23:38.212047 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d448f6767-ccmmh" event={"ID":"69be1bb2-b1e2-41fb-991c-c6a12db18fb8","Type":"ContainerStarted","Data":"6e562d01c8f9ec64152cb28f341e02ab8b11fdbc7be818ac5dd79d8ae169bf57"} Dec 02 13:23:38 crc kubenswrapper[4725]: I1202 13:23:38.212071 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6d448f6767-ccmmh" event={"ID":"69be1bb2-b1e2-41fb-991c-c6a12db18fb8","Type":"ContainerStarted","Data":"8384913fd8523a85d2171dbb65482934a36328cfc074aa71f8e70c78d9068d00"} Dec 02 13:23:38 crc kubenswrapper[4725]: I1202 13:23:38.212102 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:38 crc kubenswrapper[4725]: I1202 13:23:38.232258 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6d448f6767-ccmmh" podStartSLOduration=2.232228376 podStartE2EDuration="2.232228376s" podCreationTimestamp="2025-12-02 13:23:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:38.230527844 +0000 UTC m=+1149.187169549" watchObservedRunningTime="2025-12-02 13:23:38.232228376 +0000 UTC m=+1149.188870071" Dec 02 13:23:39 crc kubenswrapper[4725]: I1202 13:23:39.225545 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:40 crc kubenswrapper[4725]: I1202 13:23:40.358077 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:40 crc kubenswrapper[4725]: I1202 13:23:40.358653 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-central-agent" containerID="cri-o://d4729d1ae15d6e2d3424c9aee5febe575496578a32db695a550e2e92d73102a1" gracePeriod=30 Dec 02 13:23:40 crc kubenswrapper[4725]: I1202 13:23:40.358772 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-notification-agent" containerID="cri-o://e6a7f1038a30b5c4e15914bb4cf0a28df74bb6479e0b93e7663f26a6ebb615c2" gracePeriod=30 Dec 02 13:23:40 crc kubenswrapper[4725]: I1202 13:23:40.358788 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="sg-core" containerID="cri-o://8621851f97560df512e8694c05b8af2faa32a78deb1c32b54ad0ab978dc58a44" gracePeriod=30 Dec 02 13:23:40 crc kubenswrapper[4725]: I1202 13:23:40.358976 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="proxy-httpd" containerID="cri-o://905449c49700f8ba5cf3d89838967027a8981633830d93c4b1abf9ae14044138" gracePeriod=30 Dec 02 13:23:40 crc kubenswrapper[4725]: I1202 13:23:40.379492 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": EOF" Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.252292 4725 generic.go:334] "Generic (PLEG): container finished" podID="974e23a0-ac9d-419d-825e-b707b3808571" containerID="905449c49700f8ba5cf3d89838967027a8981633830d93c4b1abf9ae14044138" exitCode=0 Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.252656 4725 generic.go:334] "Generic (PLEG): container finished" podID="974e23a0-ac9d-419d-825e-b707b3808571" containerID="8621851f97560df512e8694c05b8af2faa32a78deb1c32b54ad0ab978dc58a44" exitCode=2 Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.252375 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerDied","Data":"905449c49700f8ba5cf3d89838967027a8981633830d93c4b1abf9ae14044138"} Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.252703 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerDied","Data":"8621851f97560df512e8694c05b8af2faa32a78deb1c32b54ad0ab978dc58a44"} Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.252718 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerDied","Data":"d4729d1ae15d6e2d3424c9aee5febe575496578a32db695a550e2e92d73102a1"} Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.252672 4725 generic.go:334] "Generic (PLEG): container finished" podID="974e23a0-ac9d-419d-825e-b707b3808571" containerID="d4729d1ae15d6e2d3424c9aee5febe575496578a32db695a550e2e92d73102a1" exitCode=0 Dec 02 13:23:41 crc kubenswrapper[4725]: I1202 13:23:41.793308 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 02 13:23:42 crc kubenswrapper[4725]: I1202 13:23:42.276219 4725 generic.go:334] "Generic (PLEG): container finished" podID="974e23a0-ac9d-419d-825e-b707b3808571" containerID="e6a7f1038a30b5c4e15914bb4cf0a28df74bb6479e0b93e7663f26a6ebb615c2" exitCode=0 Dec 02 13:23:42 crc kubenswrapper[4725]: I1202 13:23:42.276267 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerDied","Data":"e6a7f1038a30b5c4e15914bb4cf0a28df74bb6479e0b93e7663f26a6ebb615c2"} Dec 02 13:23:43 crc kubenswrapper[4725]: I1202 13:23:43.969672 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": dial tcp 10.217.0.164:3000: connect: connection refused" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.479598 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.482698 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6d448f6767-ccmmh" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.624089 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7f44bd87bd-2w6jw" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.753787 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.827585 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-scripts\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.827751 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-log-httpd\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.827827 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-sg-core-conf-yaml\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.827885 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-config-data\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.827963 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-run-httpd\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.828000 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nhvj\" (UniqueName: \"kubernetes.io/projected/974e23a0-ac9d-419d-825e-b707b3808571-kube-api-access-2nhvj\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.828066 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-combined-ca-bundle\") pod \"974e23a0-ac9d-419d-825e-b707b3808571\" (UID: \"974e23a0-ac9d-419d-825e-b707b3808571\") " Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.828964 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.829034 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.833549 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-scripts" (OuterVolumeSpecName: "scripts") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.839733 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/974e23a0-ac9d-419d-825e-b707b3808571-kube-api-access-2nhvj" (OuterVolumeSpecName: "kube-api-access-2nhvj") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "kube-api-access-2nhvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.865221 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.925309 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.931615 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.931871 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.931953 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.932038 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/974e23a0-ac9d-419d-825e-b707b3808571-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.932109 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nhvj\" (UniqueName: \"kubernetes.io/projected/974e23a0-ac9d-419d-825e-b707b3808571-kube-api-access-2nhvj\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.932229 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:46 crc kubenswrapper[4725]: I1202 13:23:46.936889 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-config-data" (OuterVolumeSpecName: "config-data") pod "974e23a0-ac9d-419d-825e-b707b3808571" (UID: "974e23a0-ac9d-419d-825e-b707b3808571"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.033878 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974e23a0-ac9d-419d-825e-b707b3808571-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.340140 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6ed64991-e56a-4635-b062-f68f7f1d134d","Type":"ContainerStarted","Data":"c9f03da382c6e71737353eac0253445fbfa9ac5e6a1602d3fe6272aed269861b"} Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.346138 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.346567 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"974e23a0-ac9d-419d-825e-b707b3808571","Type":"ContainerDied","Data":"427e59d400265436baf265e774250bb5d4ae43e81f0f87bdd21d0b3cca9f7386"} Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.346662 4725 scope.go:117] "RemoveContainer" containerID="905449c49700f8ba5cf3d89838967027a8981633830d93c4b1abf9ae14044138" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.371158 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.406043387 podStartE2EDuration="19.371138756s" podCreationTimestamp="2025-12-02 13:23:28 +0000 UTC" firstStartedPulling="2025-12-02 13:23:29.441586973 +0000 UTC m=+1140.398228668" lastFinishedPulling="2025-12-02 13:23:46.406682342 +0000 UTC m=+1157.363324037" observedRunningTime="2025-12-02 13:23:47.358552751 +0000 UTC m=+1158.315194446" watchObservedRunningTime="2025-12-02 13:23:47.371138756 +0000 UTC m=+1158.327780451" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.409835 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.418263 4725 scope.go:117] "RemoveContainer" containerID="8621851f97560df512e8694c05b8af2faa32a78deb1c32b54ad0ab978dc58a44" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.427314 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.449560 4725 scope.go:117] "RemoveContainer" containerID="e6a7f1038a30b5c4e15914bb4cf0a28df74bb6479e0b93e7663f26a6ebb615c2" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.451973 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:47 crc kubenswrapper[4725]: E1202 13:23:47.452454 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-notification-agent" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.452933 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-notification-agent" Dec 02 13:23:47 crc kubenswrapper[4725]: E1202 13:23:47.452952 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="proxy-httpd" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.452959 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="proxy-httpd" Dec 02 13:23:47 crc kubenswrapper[4725]: E1202 13:23:47.452974 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-central-agent" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.452980 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-central-agent" Dec 02 13:23:47 crc kubenswrapper[4725]: E1202 13:23:47.453005 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="sg-core" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.453011 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="sg-core" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.453193 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-central-agent" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.453205 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="ceilometer-notification-agent" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.453215 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="sg-core" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.453231 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="974e23a0-ac9d-419d-825e-b707b3808571" containerName="proxy-httpd" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.455099 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.457365 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.465376 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.470750 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.476836 4725 scope.go:117] "RemoveContainer" containerID="d4729d1ae15d6e2d3424c9aee5febe575496578a32db695a550e2e92d73102a1" Dec 02 13:23:47 crc kubenswrapper[4725]: E1202 13:23:47.525945 4725 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod974e23a0_ac9d_419d_825e_b707b3808571.slice\": RecentStats: unable to find data in memory cache]" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.651781 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.651847 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-run-httpd\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.652215 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-config-data\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.652293 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.652391 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-scripts\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.652474 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-log-httpd\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.652576 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdrgl\" (UniqueName: \"kubernetes.io/projected/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-kube-api-access-vdrgl\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.753957 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.754306 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-scripts\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.754346 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-log-httpd\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.754396 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdrgl\" (UniqueName: \"kubernetes.io/projected/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-kube-api-access-vdrgl\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.754478 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.754515 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-run-httpd\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.754683 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-config-data\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.755182 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-run-httpd\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.755308 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-log-httpd\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.759688 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-scripts\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.761738 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-config-data\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.763341 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.767734 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.774101 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdrgl\" (UniqueName: \"kubernetes.io/projected/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-kube-api-access-vdrgl\") pod \"ceilometer-0\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " pod="openstack/ceilometer-0" Dec 02 13:23:47 crc kubenswrapper[4725]: I1202 13:23:47.775709 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:48 crc kubenswrapper[4725]: I1202 13:23:48.233132 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:48 crc kubenswrapper[4725]: I1202 13:23:48.355393 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerStarted","Data":"5040198eab5314a5bfdc309a32aee6335ceb69f50c6d57db24823fd3bf5d8754"} Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.286532 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="974e23a0-ac9d-419d-825e-b707b3808571" path="/var/lib/kubelet/pods/974e23a0-ac9d-419d-825e-b707b3808571/volumes" Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.366486 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerStarted","Data":"163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88"} Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.627336 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.819054 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-bmtlh"] Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.820563 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.841082 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bmtlh"] Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.938643 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-qrnph"] Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.943731 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.985542 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b3bb-account-create-update-hjg7w"] Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.987014 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.990821 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 02 13:23:49 crc kubenswrapper[4725]: I1202 13:23:49.999220 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qrnph"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.005653 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww8s6\" (UniqueName: \"kubernetes.io/projected/d35ce79f-44d1-4180-a786-32dd8df46b76-kube-api-access-ww8s6\") pod \"nova-api-db-create-bmtlh\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.005826 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d35ce79f-44d1-4180-a786-32dd8df46b76-operator-scripts\") pod \"nova-api-db-create-bmtlh\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.024564 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b3bb-account-create-update-hjg7w"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.112668 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d35ce79f-44d1-4180-a786-32dd8df46b76-operator-scripts\") pod \"nova-api-db-create-bmtlh\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.112781 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww8s6\" (UniqueName: \"kubernetes.io/projected/d35ce79f-44d1-4180-a786-32dd8df46b76-kube-api-access-ww8s6\") pod \"nova-api-db-create-bmtlh\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.112818 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e26ab0f-08f5-4712-a40d-ff08b563b752-operator-scripts\") pod \"nova-api-b3bb-account-create-update-hjg7w\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.112882 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fvbk\" (UniqueName: \"kubernetes.io/projected/1e26ab0f-08f5-4712-a40d-ff08b563b752-kube-api-access-2fvbk\") pod \"nova-api-b3bb-account-create-update-hjg7w\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.112919 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-operator-scripts\") pod \"nova-cell0-db-create-qrnph\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.112973 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nxmd\" (UniqueName: \"kubernetes.io/projected/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-kube-api-access-5nxmd\") pod \"nova-cell0-db-create-qrnph\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.114363 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d35ce79f-44d1-4180-a786-32dd8df46b76-operator-scripts\") pod \"nova-api-db-create-bmtlh\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.136691 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xkn8g"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.138098 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.164145 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xkn8g"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.168213 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww8s6\" (UniqueName: \"kubernetes.io/projected/d35ce79f-44d1-4180-a786-32dd8df46b76-kube-api-access-ww8s6\") pod \"nova-api-db-create-bmtlh\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.215641 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e26ab0f-08f5-4712-a40d-ff08b563b752-operator-scripts\") pod \"nova-api-b3bb-account-create-update-hjg7w\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.215718 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fvbk\" (UniqueName: \"kubernetes.io/projected/1e26ab0f-08f5-4712-a40d-ff08b563b752-kube-api-access-2fvbk\") pod \"nova-api-b3bb-account-create-update-hjg7w\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.215752 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-operator-scripts\") pod \"nova-cell0-db-create-qrnph\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.215806 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nxmd\" (UniqueName: \"kubernetes.io/projected/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-kube-api-access-5nxmd\") pod \"nova-cell0-db-create-qrnph\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.216890 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e26ab0f-08f5-4712-a40d-ff08b563b752-operator-scripts\") pod \"nova-api-b3bb-account-create-update-hjg7w\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.217244 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-operator-scripts\") pod \"nova-cell0-db-create-qrnph\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.227303 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-80a5-account-create-update-4pzhq"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.228964 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.232024 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.233930 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fvbk\" (UniqueName: \"kubernetes.io/projected/1e26ab0f-08f5-4712-a40d-ff08b563b752-kube-api-access-2fvbk\") pod \"nova-api-b3bb-account-create-update-hjg7w\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.234077 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nxmd\" (UniqueName: \"kubernetes.io/projected/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-kube-api-access-5nxmd\") pod \"nova-cell0-db-create-qrnph\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.260152 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-80a5-account-create-update-4pzhq"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.289973 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.318588 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46a0d2b3-489c-4160-a7c0-7570d7baca42-operator-scripts\") pod \"nova-cell1-db-create-xkn8g\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.318677 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhmjr\" (UniqueName: \"kubernetes.io/projected/46a0d2b3-489c-4160-a7c0-7570d7baca42-kube-api-access-zhmjr\") pod \"nova-cell1-db-create-xkn8g\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.318720 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlf6f\" (UniqueName: \"kubernetes.io/projected/2a1f9729-56c5-4955-86b3-68f05bb15396-kube-api-access-nlf6f\") pod \"nova-cell0-80a5-account-create-update-4pzhq\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.318982 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.319120 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a1f9729-56c5-4955-86b3-68f05bb15396-operator-scripts\") pod \"nova-cell0-80a5-account-create-update-4pzhq\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.319622 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-79dd-account-create-update-h92mv"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.321132 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.324815 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.339138 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-79dd-account-create-update-h92mv"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.392201 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerStarted","Data":"e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5"} Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.420306 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0d25cc-d587-4265-b6d3-373ea0e46728-operator-scripts\") pod \"nova-cell1-79dd-account-create-update-h92mv\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.420965 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46a0d2b3-489c-4160-a7c0-7570d7baca42-operator-scripts\") pod \"nova-cell1-db-create-xkn8g\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.421014 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhmjr\" (UniqueName: \"kubernetes.io/projected/46a0d2b3-489c-4160-a7c0-7570d7baca42-kube-api-access-zhmjr\") pod \"nova-cell1-db-create-xkn8g\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.421051 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlf6f\" (UniqueName: \"kubernetes.io/projected/2a1f9729-56c5-4955-86b3-68f05bb15396-kube-api-access-nlf6f\") pod \"nova-cell0-80a5-account-create-update-4pzhq\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.421106 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc6rs\" (UniqueName: \"kubernetes.io/projected/5c0d25cc-d587-4265-b6d3-373ea0e46728-kube-api-access-hc6rs\") pod \"nova-cell1-79dd-account-create-update-h92mv\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.421151 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a1f9729-56c5-4955-86b3-68f05bb15396-operator-scripts\") pod \"nova-cell0-80a5-account-create-update-4pzhq\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.422085 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a1f9729-56c5-4955-86b3-68f05bb15396-operator-scripts\") pod \"nova-cell0-80a5-account-create-update-4pzhq\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.422587 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46a0d2b3-489c-4160-a7c0-7570d7baca42-operator-scripts\") pod \"nova-cell1-db-create-xkn8g\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.447966 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlf6f\" (UniqueName: \"kubernetes.io/projected/2a1f9729-56c5-4955-86b3-68f05bb15396-kube-api-access-nlf6f\") pod \"nova-cell0-80a5-account-create-update-4pzhq\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.452907 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhmjr\" (UniqueName: \"kubernetes.io/projected/46a0d2b3-489c-4160-a7c0-7570d7baca42-kube-api-access-zhmjr\") pod \"nova-cell1-db-create-xkn8g\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.468661 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.511568 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.522475 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc6rs\" (UniqueName: \"kubernetes.io/projected/5c0d25cc-d587-4265-b6d3-373ea0e46728-kube-api-access-hc6rs\") pod \"nova-cell1-79dd-account-create-update-h92mv\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.522594 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0d25cc-d587-4265-b6d3-373ea0e46728-operator-scripts\") pod \"nova-cell1-79dd-account-create-update-h92mv\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.523783 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0d25cc-d587-4265-b6d3-373ea0e46728-operator-scripts\") pod \"nova-cell1-79dd-account-create-update-h92mv\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.542226 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc6rs\" (UniqueName: \"kubernetes.io/projected/5c0d25cc-d587-4265-b6d3-373ea0e46728-kube-api-access-hc6rs\") pod \"nova-cell1-79dd-account-create-update-h92mv\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.593777 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.719573 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.896512 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b3bb-account-create-update-hjg7w"] Dec 02 13:23:50 crc kubenswrapper[4725]: I1202 13:23:50.907484 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qrnph"] Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.309585 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bmtlh"] Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.339771 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-79dd-account-create-update-h92mv"] Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.351321 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-80a5-account-create-update-4pzhq"] Dec 02 13:23:51 crc kubenswrapper[4725]: W1202 13:23:51.358194 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a1f9729_56c5_4955_86b3_68f05bb15396.slice/crio-640f7a8bffcfb33cd9d0ee4e09cda83060dc29913f10473e3daadfb8897a697d WatchSource:0}: Error finding container 640f7a8bffcfb33cd9d0ee4e09cda83060dc29913f10473e3daadfb8897a697d: Status 404 returned error can't find the container with id 640f7a8bffcfb33cd9d0ee4e09cda83060dc29913f10473e3daadfb8897a697d Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.360618 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xkn8g"] Dec 02 13:23:51 crc kubenswrapper[4725]: W1202 13:23:51.373436 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46a0d2b3_489c_4160_a7c0_7570d7baca42.slice/crio-326e79986c73d99385d0a4dbacd41ab1856e7798244ceaaaeaf8407e3db05957 WatchSource:0}: Error finding container 326e79986c73d99385d0a4dbacd41ab1856e7798244ceaaaeaf8407e3db05957: Status 404 returned error can't find the container with id 326e79986c73d99385d0a4dbacd41ab1856e7798244ceaaaeaf8407e3db05957 Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.449604 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmtlh" event={"ID":"d35ce79f-44d1-4180-a786-32dd8df46b76","Type":"ContainerStarted","Data":"50f4020b1fe544a9b51458fc2823329d9e5bcb1befb2b3d9824329649f5b3d29"} Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.451445 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" event={"ID":"2a1f9729-56c5-4955-86b3-68f05bb15396","Type":"ContainerStarted","Data":"640f7a8bffcfb33cd9d0ee4e09cda83060dc29913f10473e3daadfb8897a697d"} Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.453831 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qrnph" event={"ID":"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53","Type":"ContainerStarted","Data":"46a44150c2deb764c12f87184c8da8030ddc7df17b8c1b63f8e6a57fc8af0924"} Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.455915 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" event={"ID":"1e26ab0f-08f5-4712-a40d-ff08b563b752","Type":"ContainerStarted","Data":"f7c6d706ae70b7159165777379e01e4a52b5989bacbb60f0afce33219905ebe7"} Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.458620 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xkn8g" event={"ID":"46a0d2b3-489c-4160-a7c0-7570d7baca42","Type":"ContainerStarted","Data":"326e79986c73d99385d0a4dbacd41ab1856e7798244ceaaaeaf8407e3db05957"} Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.465321 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerStarted","Data":"e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4"} Dec 02 13:23:51 crc kubenswrapper[4725]: I1202 13:23:51.471845 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" event={"ID":"5c0d25cc-d587-4265-b6d3-373ea0e46728","Type":"ContainerStarted","Data":"7598c02366286d5e2a6bae4c1bb696d3b7ae344c00d2a3d161d72fa9cd496735"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.432843 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.438536 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-log" containerID="cri-o://400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e" gracePeriod=30 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.438703 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-httpd" containerID="cri-o://5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75" gracePeriod=30 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.499724 4725 generic.go:334] "Generic (PLEG): container finished" podID="1e26ab0f-08f5-4712-a40d-ff08b563b752" containerID="0811559d7ba7397b452caaff96197756e30abbfff1f08de1ac4ff06212495fa9" exitCode=0 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.499789 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" event={"ID":"1e26ab0f-08f5-4712-a40d-ff08b563b752","Type":"ContainerDied","Data":"0811559d7ba7397b452caaff96197756e30abbfff1f08de1ac4ff06212495fa9"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.503326 4725 generic.go:334] "Generic (PLEG): container finished" podID="46a0d2b3-489c-4160-a7c0-7570d7baca42" containerID="a736568863b41adcae7a1f3fbe454c3b1b0a859bb9d14317860e4451a802abc2" exitCode=0 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.503399 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xkn8g" event={"ID":"46a0d2b3-489c-4160-a7c0-7570d7baca42","Type":"ContainerDied","Data":"a736568863b41adcae7a1f3fbe454c3b1b0a859bb9d14317860e4451a802abc2"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.515757 4725 generic.go:334] "Generic (PLEG): container finished" podID="5c0d25cc-d587-4265-b6d3-373ea0e46728" containerID="e2dcb35682ba1ded69f2dbde8efad46005747c258d77b01333bdb79e3dbcd36f" exitCode=0 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.516180 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" event={"ID":"5c0d25cc-d587-4265-b6d3-373ea0e46728","Type":"ContainerDied","Data":"e2dcb35682ba1ded69f2dbde8efad46005747c258d77b01333bdb79e3dbcd36f"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.539622 4725 generic.go:334] "Generic (PLEG): container finished" podID="d35ce79f-44d1-4180-a786-32dd8df46b76" containerID="6e46b5e899fd2cf06fb8a020444ddc795c662bf252af91c3b2cad06068472703" exitCode=0 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.539692 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmtlh" event={"ID":"d35ce79f-44d1-4180-a786-32dd8df46b76","Type":"ContainerDied","Data":"6e46b5e899fd2cf06fb8a020444ddc795c662bf252af91c3b2cad06068472703"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.547516 4725 generic.go:334] "Generic (PLEG): container finished" podID="2a1f9729-56c5-4955-86b3-68f05bb15396" containerID="a64a2dc372541bc13f6ceef47b73d7c01580d1c10a6a54d32eed66cbe9ff2e49" exitCode=0 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.547606 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" event={"ID":"2a1f9729-56c5-4955-86b3-68f05bb15396","Type":"ContainerDied","Data":"a64a2dc372541bc13f6ceef47b73d7c01580d1c10a6a54d32eed66cbe9ff2e49"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.551339 4725 generic.go:334] "Generic (PLEG): container finished" podID="1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" containerID="cae6e01996efca918308c0ea83db409419a9d0c9b4ec7ddadc172962a6884fa7" exitCode=0 Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.551393 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qrnph" event={"ID":"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53","Type":"ContainerDied","Data":"cae6e01996efca918308c0ea83db409419a9d0c9b4ec7ddadc172962a6884fa7"} Dec 02 13:23:52 crc kubenswrapper[4725]: I1202 13:23:52.935124 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.026994 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kftlz\" (UniqueName: \"kubernetes.io/projected/cd102f85-1b92-4215-95e9-33baf5949dec-kube-api-access-kftlz\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027073 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-config-data\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027143 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-secret-key\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027179 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd102f85-1b92-4215-95e9-33baf5949dec-logs\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027202 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-combined-ca-bundle\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027295 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-scripts\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027370 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-tls-certs\") pod \"cd102f85-1b92-4215-95e9-33baf5949dec\" (UID: \"cd102f85-1b92-4215-95e9-33baf5949dec\") " Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.027624 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd102f85-1b92-4215-95e9-33baf5949dec-logs" (OuterVolumeSpecName: "logs") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.028177 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd102f85-1b92-4215-95e9-33baf5949dec-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.034702 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd102f85-1b92-4215-95e9-33baf5949dec-kube-api-access-kftlz" (OuterVolumeSpecName: "kube-api-access-kftlz") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "kube-api-access-kftlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.035732 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.062086 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-scripts" (OuterVolumeSpecName: "scripts") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.064647 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.073861 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-config-data" (OuterVolumeSpecName: "config-data") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.114876 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "cd102f85-1b92-4215-95e9-33baf5949dec" (UID: "cd102f85-1b92-4215-95e9-33baf5949dec"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.139625 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kftlz\" (UniqueName: \"kubernetes.io/projected/cd102f85-1b92-4215-95e9-33baf5949dec-kube-api-access-kftlz\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.140083 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.140097 4725 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.140108 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.140121 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd102f85-1b92-4215-95e9-33baf5949dec-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.140132 4725 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd102f85-1b92-4215-95e9-33baf5949dec-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.564904 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerStarted","Data":"6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f"} Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.565036 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-central-agent" containerID="cri-o://163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88" gracePeriod=30 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.565086 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.565093 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="proxy-httpd" containerID="cri-o://6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f" gracePeriod=30 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.565147 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-notification-agent" containerID="cri-o://e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5" gracePeriod=30 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.565133 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="sg-core" containerID="cri-o://e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4" gracePeriod=30 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.583277 4725 generic.go:334] "Generic (PLEG): container finished" podID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerID="400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e" exitCode=143 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.583390 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29ba5625-386c-4e5f-8aa1-7ed349af1d7a","Type":"ContainerDied","Data":"400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e"} Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.587415 4725 generic.go:334] "Generic (PLEG): container finished" podID="cd102f85-1b92-4215-95e9-33baf5949dec" containerID="149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077" exitCode=137 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.587611 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44bd87bd-2w6jw" event={"ID":"cd102f85-1b92-4215-95e9-33baf5949dec","Type":"ContainerDied","Data":"149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077"} Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.587673 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44bd87bd-2w6jw" event={"ID":"cd102f85-1b92-4215-95e9-33baf5949dec","Type":"ContainerDied","Data":"c9f9992cd5ffd2a3837f9f5d8a78714847a99fd4af568dd4010844d47566a212"} Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.587695 4725 scope.go:117] "RemoveContainer" containerID="0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.587765 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44bd87bd-2w6jw" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.669295 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.669857 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-log" containerID="cri-o://6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6" gracePeriod=30 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.670405 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-httpd" containerID="cri-o://568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9" gracePeriod=30 Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.706405 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5789473640000002 podStartE2EDuration="6.706375357s" podCreationTimestamp="2025-12-02 13:23:47 +0000 UTC" firstStartedPulling="2025-12-02 13:23:48.244370248 +0000 UTC m=+1159.201011943" lastFinishedPulling="2025-12-02 13:23:52.371798241 +0000 UTC m=+1163.328439936" observedRunningTime="2025-12-02 13:23:53.608968469 +0000 UTC m=+1164.565610164" watchObservedRunningTime="2025-12-02 13:23:53.706375357 +0000 UTC m=+1164.663017042" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.733285 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f44bd87bd-2w6jw"] Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.753539 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f44bd87bd-2w6jw"] Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.874950 4725 scope.go:117] "RemoveContainer" containerID="149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.988992 4725 scope.go:117] "RemoveContainer" containerID="0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360" Dec 02 13:23:53 crc kubenswrapper[4725]: E1202 13:23:53.989656 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360\": container with ID starting with 0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360 not found: ID does not exist" containerID="0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.989684 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360"} err="failed to get container status \"0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360\": rpc error: code = NotFound desc = could not find container \"0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360\": container with ID starting with 0a3ae35990c5646a5ad132e1228e1f5a36a433b4a7aa841f413214bfdd54f360 not found: ID does not exist" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.989703 4725 scope.go:117] "RemoveContainer" containerID="149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077" Dec 02 13:23:53 crc kubenswrapper[4725]: E1202 13:23:53.990717 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077\": container with ID starting with 149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077 not found: ID does not exist" containerID="149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077" Dec 02 13:23:53 crc kubenswrapper[4725]: I1202 13:23:53.990751 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077"} err="failed to get container status \"149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077\": rpc error: code = NotFound desc = could not find container \"149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077\": container with ID starting with 149e1f751217f67d5ca8e20bb3ed3d5454a84173618e799421f500a6e2aac077 not found: ID does not exist" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.152344 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.160453 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhmjr\" (UniqueName: \"kubernetes.io/projected/46a0d2b3-489c-4160-a7c0-7570d7baca42-kube-api-access-zhmjr\") pod \"46a0d2b3-489c-4160-a7c0-7570d7baca42\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.160524 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46a0d2b3-489c-4160-a7c0-7570d7baca42-operator-scripts\") pod \"46a0d2b3-489c-4160-a7c0-7570d7baca42\" (UID: \"46a0d2b3-489c-4160-a7c0-7570d7baca42\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.161798 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46a0d2b3-489c-4160-a7c0-7570d7baca42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "46a0d2b3-489c-4160-a7c0-7570d7baca42" (UID: "46a0d2b3-489c-4160-a7c0-7570d7baca42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.168155 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a0d2b3-489c-4160-a7c0-7570d7baca42-kube-api-access-zhmjr" (OuterVolumeSpecName: "kube-api-access-zhmjr") pod "46a0d2b3-489c-4160-a7c0-7570d7baca42" (UID: "46a0d2b3-489c-4160-a7c0-7570d7baca42"). InnerVolumeSpecName "kube-api-access-zhmjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.262485 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhmjr\" (UniqueName: \"kubernetes.io/projected/46a0d2b3-489c-4160-a7c0-7570d7baca42-kube-api-access-zhmjr\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.262533 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/46a0d2b3-489c-4160-a7c0-7570d7baca42-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.287870 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.287921 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.287966 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.288706 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cf7371c50f3e4f7f952b1130b75faf18d2e6023f1c852348322f64d11754fac"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.288757 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://1cf7371c50f3e4f7f952b1130b75faf18d2e6023f1c852348322f64d11754fac" gracePeriod=600 Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.377422 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.383514 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.404714 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.415065 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.429284 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570106 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nxmd\" (UniqueName: \"kubernetes.io/projected/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-kube-api-access-5nxmd\") pod \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570203 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fvbk\" (UniqueName: \"kubernetes.io/projected/1e26ab0f-08f5-4712-a40d-ff08b563b752-kube-api-access-2fvbk\") pod \"1e26ab0f-08f5-4712-a40d-ff08b563b752\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570342 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0d25cc-d587-4265-b6d3-373ea0e46728-operator-scripts\") pod \"5c0d25cc-d587-4265-b6d3-373ea0e46728\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570438 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc6rs\" (UniqueName: \"kubernetes.io/projected/5c0d25cc-d587-4265-b6d3-373ea0e46728-kube-api-access-hc6rs\") pod \"5c0d25cc-d587-4265-b6d3-373ea0e46728\" (UID: \"5c0d25cc-d587-4265-b6d3-373ea0e46728\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570517 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a1f9729-56c5-4955-86b3-68f05bb15396-operator-scripts\") pod \"2a1f9729-56c5-4955-86b3-68f05bb15396\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570567 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e26ab0f-08f5-4712-a40d-ff08b563b752-operator-scripts\") pod \"1e26ab0f-08f5-4712-a40d-ff08b563b752\" (UID: \"1e26ab0f-08f5-4712-a40d-ff08b563b752\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570619 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlf6f\" (UniqueName: \"kubernetes.io/projected/2a1f9729-56c5-4955-86b3-68f05bb15396-kube-api-access-nlf6f\") pod \"2a1f9729-56c5-4955-86b3-68f05bb15396\" (UID: \"2a1f9729-56c5-4955-86b3-68f05bb15396\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570653 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww8s6\" (UniqueName: \"kubernetes.io/projected/d35ce79f-44d1-4180-a786-32dd8df46b76-kube-api-access-ww8s6\") pod \"d35ce79f-44d1-4180-a786-32dd8df46b76\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570676 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-operator-scripts\") pod \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\" (UID: \"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.570712 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d35ce79f-44d1-4180-a786-32dd8df46b76-operator-scripts\") pod \"d35ce79f-44d1-4180-a786-32dd8df46b76\" (UID: \"d35ce79f-44d1-4180-a786-32dd8df46b76\") " Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.571727 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d35ce79f-44d1-4180-a786-32dd8df46b76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d35ce79f-44d1-4180-a786-32dd8df46b76" (UID: "d35ce79f-44d1-4180-a786-32dd8df46b76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.572400 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e26ab0f-08f5-4712-a40d-ff08b563b752-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e26ab0f-08f5-4712-a40d-ff08b563b752" (UID: "1e26ab0f-08f5-4712-a40d-ff08b563b752"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.572564 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" (UID: "1a9ce041-5d7a-4a5f-be52-fdfae2f06a53"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.572567 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0d25cc-d587-4265-b6d3-373ea0e46728-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c0d25cc-d587-4265-b6d3-373ea0e46728" (UID: "5c0d25cc-d587-4265-b6d3-373ea0e46728"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.572640 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a1f9729-56c5-4955-86b3-68f05bb15396-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a1f9729-56c5-4955-86b3-68f05bb15396" (UID: "2a1f9729-56c5-4955-86b3-68f05bb15396"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.576752 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-kube-api-access-5nxmd" (OuterVolumeSpecName: "kube-api-access-5nxmd") pod "1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" (UID: "1a9ce041-5d7a-4a5f-be52-fdfae2f06a53"). InnerVolumeSpecName "kube-api-access-5nxmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.576807 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e26ab0f-08f5-4712-a40d-ff08b563b752-kube-api-access-2fvbk" (OuterVolumeSpecName: "kube-api-access-2fvbk") pod "1e26ab0f-08f5-4712-a40d-ff08b563b752" (UID: "1e26ab0f-08f5-4712-a40d-ff08b563b752"). InnerVolumeSpecName "kube-api-access-2fvbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.580193 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a1f9729-56c5-4955-86b3-68f05bb15396-kube-api-access-nlf6f" (OuterVolumeSpecName: "kube-api-access-nlf6f") pod "2a1f9729-56c5-4955-86b3-68f05bb15396" (UID: "2a1f9729-56c5-4955-86b3-68f05bb15396"). InnerVolumeSpecName "kube-api-access-nlf6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.580328 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0d25cc-d587-4265-b6d3-373ea0e46728-kube-api-access-hc6rs" (OuterVolumeSpecName: "kube-api-access-hc6rs") pod "5c0d25cc-d587-4265-b6d3-373ea0e46728" (UID: "5c0d25cc-d587-4265-b6d3-373ea0e46728"). InnerVolumeSpecName "kube-api-access-hc6rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.587287 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d35ce79f-44d1-4180-a786-32dd8df46b76-kube-api-access-ww8s6" (OuterVolumeSpecName: "kube-api-access-ww8s6") pod "d35ce79f-44d1-4180-a786-32dd8df46b76" (UID: "d35ce79f-44d1-4180-a786-32dd8df46b76"). InnerVolumeSpecName "kube-api-access-ww8s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.603513 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.603856 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-80a5-account-create-update-4pzhq" event={"ID":"2a1f9729-56c5-4955-86b3-68f05bb15396","Type":"ContainerDied","Data":"640f7a8bffcfb33cd9d0ee4e09cda83060dc29913f10473e3daadfb8897a697d"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.603935 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="640f7a8bffcfb33cd9d0ee4e09cda83060dc29913f10473e3daadfb8897a697d" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.606529 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xkn8g" event={"ID":"46a0d2b3-489c-4160-a7c0-7570d7baca42","Type":"ContainerDied","Data":"326e79986c73d99385d0a4dbacd41ab1856e7798244ceaaaeaf8407e3db05957"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.606561 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="326e79986c73d99385d0a4dbacd41ab1856e7798244ceaaaeaf8407e3db05957" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.606630 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xkn8g" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.615332 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" event={"ID":"5c0d25cc-d587-4265-b6d3-373ea0e46728","Type":"ContainerDied","Data":"7598c02366286d5e2a6bae4c1bb696d3b7ae344c00d2a3d161d72fa9cd496735"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.615397 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7598c02366286d5e2a6bae4c1bb696d3b7ae344c00d2a3d161d72fa9cd496735" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.615491 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-79dd-account-create-update-h92mv" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.619069 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="1cf7371c50f3e4f7f952b1130b75faf18d2e6023f1c852348322f64d11754fac" exitCode=0 Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.619160 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"1cf7371c50f3e4f7f952b1130b75faf18d2e6023f1c852348322f64d11754fac"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.619208 4725 scope.go:117] "RemoveContainer" containerID="203c78b7d89a690dfa6df4df74f2999e81c6261249dd7853e1b1c46205cb6197" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.622193 4725 generic.go:334] "Generic (PLEG): container finished" podID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerID="6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6" exitCode=143 Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.622264 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"153f44c9-3565-45bc-a162-8b7646df9cb2","Type":"ContainerDied","Data":"6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.625148 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qrnph" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.625137 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qrnph" event={"ID":"1a9ce041-5d7a-4a5f-be52-fdfae2f06a53","Type":"ContainerDied","Data":"46a44150c2deb764c12f87184c8da8030ddc7df17b8c1b63f8e6a57fc8af0924"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.625291 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46a44150c2deb764c12f87184c8da8030ddc7df17b8c1b63f8e6a57fc8af0924" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.631310 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" event={"ID":"1e26ab0f-08f5-4712-a40d-ff08b563b752","Type":"ContainerDied","Data":"f7c6d706ae70b7159165777379e01e4a52b5989bacbb60f0afce33219905ebe7"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.631349 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7c6d706ae70b7159165777379e01e4a52b5989bacbb60f0afce33219905ebe7" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.631356 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b3bb-account-create-update-hjg7w" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.639626 4725 generic.go:334] "Generic (PLEG): container finished" podID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerID="6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f" exitCode=0 Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.639663 4725 generic.go:334] "Generic (PLEG): container finished" podID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerID="e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4" exitCode=2 Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.639672 4725 generic.go:334] "Generic (PLEG): container finished" podID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerID="e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5" exitCode=0 Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.639720 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerDied","Data":"6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.639748 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerDied","Data":"e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.639762 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerDied","Data":"e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.646654 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bmtlh" event={"ID":"d35ce79f-44d1-4180-a786-32dd8df46b76","Type":"ContainerDied","Data":"50f4020b1fe544a9b51458fc2823329d9e5bcb1befb2b3d9824329649f5b3d29"} Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.646699 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50f4020b1fe544a9b51458fc2823329d9e5bcb1befb2b3d9824329649f5b3d29" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.646756 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bmtlh" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673724 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlf6f\" (UniqueName: \"kubernetes.io/projected/2a1f9729-56c5-4955-86b3-68f05bb15396-kube-api-access-nlf6f\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673757 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww8s6\" (UniqueName: \"kubernetes.io/projected/d35ce79f-44d1-4180-a786-32dd8df46b76-kube-api-access-ww8s6\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673768 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673800 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d35ce79f-44d1-4180-a786-32dd8df46b76-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673839 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nxmd\" (UniqueName: \"kubernetes.io/projected/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53-kube-api-access-5nxmd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673851 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fvbk\" (UniqueName: \"kubernetes.io/projected/1e26ab0f-08f5-4712-a40d-ff08b563b752-kube-api-access-2fvbk\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673863 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0d25cc-d587-4265-b6d3-373ea0e46728-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673875 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc6rs\" (UniqueName: \"kubernetes.io/projected/5c0d25cc-d587-4265-b6d3-373ea0e46728-kube-api-access-hc6rs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673884 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a1f9729-56c5-4955-86b3-68f05bb15396-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:54 crc kubenswrapper[4725]: I1202 13:23:54.673893 4725 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e26ab0f-08f5-4712-a40d-ff08b563b752-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:55 crc kubenswrapper[4725]: I1202 13:23:55.279483 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" path="/var/lib/kubelet/pods/cd102f85-1b92-4215-95e9-33baf5949dec/volumes" Dec 02 13:23:55 crc kubenswrapper[4725]: I1202 13:23:55.662564 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"05f4c4672c17d44bd1b7bcedcce35ce67bcd6b4893164e4887e1ed3ea23f75fe"} Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.208255 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412555 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-config-data\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412625 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-httpd-run\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412685 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-logs\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412773 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-combined-ca-bundle\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412802 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bgsc\" (UniqueName: \"kubernetes.io/projected/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-kube-api-access-2bgsc\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412862 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-scripts\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412956 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.412991 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-public-tls-certs\") pod \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\" (UID: \"29ba5625-386c-4e5f-8aa1-7ed349af1d7a\") " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.413156 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.413427 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-logs" (OuterVolumeSpecName: "logs") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.413665 4725 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.413690 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.424976 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-kube-api-access-2bgsc" (OuterVolumeSpecName: "kube-api-access-2bgsc") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "kube-api-access-2bgsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.433624 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-scripts" (OuterVolumeSpecName: "scripts") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.434712 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.483772 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-config-data" (OuterVolumeSpecName: "config-data") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.491860 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.504713 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "29ba5625-386c-4e5f-8aa1-7ed349af1d7a" (UID: "29ba5625-386c-4e5f-8aa1-7ed349af1d7a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.517936 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.517974 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.517986 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bgsc\" (UniqueName: \"kubernetes.io/projected/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-kube-api-access-2bgsc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.517995 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.518022 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.518032 4725 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29ba5625-386c-4e5f-8aa1-7ed349af1d7a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.548045 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.619624 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.673996 4725 generic.go:334] "Generic (PLEG): container finished" podID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerID="5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75" exitCode=0 Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.674120 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.674114 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29ba5625-386c-4e5f-8aa1-7ed349af1d7a","Type":"ContainerDied","Data":"5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75"} Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.675278 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"29ba5625-386c-4e5f-8aa1-7ed349af1d7a","Type":"ContainerDied","Data":"362bfba2783ee3da80abcdbf57b938ebf5cdd43fd4587798d8af88a39089b2ee"} Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.675314 4725 scope.go:117] "RemoveContainer" containerID="5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.700178 4725 scope.go:117] "RemoveContainer" containerID="400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.715107 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.722999 4725 scope.go:117] "RemoveContainer" containerID="5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.723693 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75\": container with ID starting with 5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75 not found: ID does not exist" containerID="5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.723737 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75"} err="failed to get container status \"5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75\": rpc error: code = NotFound desc = could not find container \"5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75\": container with ID starting with 5f07f7e46cb187210a7ee6c974067f96078283b95e9adb95ab29f92422a35e75 not found: ID does not exist" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.723763 4725 scope.go:117] "RemoveContainer" containerID="400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.724800 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e\": container with ID starting with 400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e not found: ID does not exist" containerID="400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.724916 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e"} err="failed to get container status \"400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e\": rpc error: code = NotFound desc = could not find container \"400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e\": container with ID starting with 400fc6523428bf8e13139bcf5dae734b2c76081ac1e6baf8846253667eebe85e not found: ID does not exist" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.728844 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.744377 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745126 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745145 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745162 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-httpd" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745169 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-httpd" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745181 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-log" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745187 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-log" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745197 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e26ab0f-08f5-4712-a40d-ff08b563b752" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745205 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e26ab0f-08f5-4712-a40d-ff08b563b752" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745217 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d35ce79f-44d1-4180-a786-32dd8df46b76" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745223 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d35ce79f-44d1-4180-a786-32dd8df46b76" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745231 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0d25cc-d587-4265-b6d3-373ea0e46728" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745238 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0d25cc-d587-4265-b6d3-373ea0e46728" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745248 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745255 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745266 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a0d2b3-489c-4160-a7c0-7570d7baca42" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745272 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a0d2b3-489c-4160-a7c0-7570d7baca42" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745288 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1f9729-56c5-4955-86b3-68f05bb15396" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745294 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1f9729-56c5-4955-86b3-68f05bb15396" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: E1202 13:23:56.745313 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon-log" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745319 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon-log" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745490 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d35ce79f-44d1-4180-a786-32dd8df46b76" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745501 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745509 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a1f9729-56c5-4955-86b3-68f05bb15396" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745520 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e26ab0f-08f5-4712-a40d-ff08b563b752" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745531 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745542 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd102f85-1b92-4215-95e9-33baf5949dec" containerName="horizon-log" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745553 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-log" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745562 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0d25cc-d587-4265-b6d3-373ea0e46728" containerName="mariadb-account-create-update" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745573 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" containerName="glance-httpd" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.745581 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a0d2b3-489c-4160-a7c0-7570d7baca42" containerName="mariadb-database-create" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.746837 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.749115 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.749325 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.758775 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.929473 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.929711 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.929848 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48s9f\" (UniqueName: \"kubernetes.io/projected/2b150790-de06-4a61-901f-014c70552a9c-kube-api-access-48s9f\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.929937 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.930200 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b150790-de06-4a61-901f-014c70552a9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.930298 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b150790-de06-4a61-901f-014c70552a9c-logs\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.930332 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:56 crc kubenswrapper[4725]: I1202 13:23:56.930394 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031752 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031793 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b150790-de06-4a61-901f-014c70552a9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031834 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b150790-de06-4a61-901f-014c70552a9c-logs\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031857 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031887 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031922 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031963 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.031998 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48s9f\" (UniqueName: \"kubernetes.io/projected/2b150790-de06-4a61-901f-014c70552a9c-kube-api-access-48s9f\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.032543 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.038068 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b150790-de06-4a61-901f-014c70552a9c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.038338 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b150790-de06-4a61-901f-014c70552a9c-logs\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.047092 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.055836 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.056479 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.058426 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48s9f\" (UniqueName: \"kubernetes.io/projected/2b150790-de06-4a61-901f-014c70552a9c-kube-api-access-48s9f\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.059368 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b150790-de06-4a61-901f-014c70552a9c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.083234 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-external-api-0\" (UID: \"2b150790-de06-4a61-901f-014c70552a9c\") " pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.279546 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29ba5625-386c-4e5f-8aa1-7ed349af1d7a" path="/var/lib/kubelet/pods/29ba5625-386c-4e5f-8aa1-7ed349af1d7a/volumes" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.379171 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.493198 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.632711 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.646589 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-log-httpd\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.646780 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-config-data\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.646868 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-scripts\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.646892 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-run-httpd\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.646956 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-combined-ca-bundle\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.646990 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdrgl\" (UniqueName: \"kubernetes.io/projected/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-kube-api-access-vdrgl\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.647036 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-sg-core-conf-yaml\") pod \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\" (UID: \"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.648847 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.649445 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.669570 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-kube-api-access-vdrgl" (OuterVolumeSpecName: "kube-api-access-vdrgl") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "kube-api-access-vdrgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.671618 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-scripts" (OuterVolumeSpecName: "scripts") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.699245 4725 generic.go:334] "Generic (PLEG): container finished" podID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerID="568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9" exitCode=0 Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.699328 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"153f44c9-3565-45bc-a162-8b7646df9cb2","Type":"ContainerDied","Data":"568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9"} Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.699355 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"153f44c9-3565-45bc-a162-8b7646df9cb2","Type":"ContainerDied","Data":"3a0bedf12eee23f5f1c996d4bfddf3fcad9ee79eb97cd73a24d1fe46a3652a9d"} Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.699372 4725 scope.go:117] "RemoveContainer" containerID="568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.701487 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.716899 4725 generic.go:334] "Generic (PLEG): container finished" podID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerID="163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88" exitCode=0 Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.716965 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerDied","Data":"163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88"} Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.716995 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06fe3c8a-ba63-42d9-ae09-3b62aadb7b21","Type":"ContainerDied","Data":"5040198eab5314a5bfdc309a32aee6335ceb69f50c6d57db24823fd3bf5d8754"} Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.717058 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.733984 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.734400 4725 scope.go:117] "RemoveContainer" containerID="6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.749567 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-httpd-run\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.749840 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccl4r\" (UniqueName: \"kubernetes.io/projected/153f44c9-3565-45bc-a162-8b7646df9cb2-kube-api-access-ccl4r\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.749982 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-scripts\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.750089 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-combined-ca-bundle\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.750203 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-config-data\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.750206 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.750338 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-internal-tls-certs\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.750785 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-logs\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.750963 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"153f44c9-3565-45bc-a162-8b7646df9cb2\" (UID: \"153f44c9-3565-45bc-a162-8b7646df9cb2\") " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.751816 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.751914 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.751992 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdrgl\" (UniqueName: \"kubernetes.io/projected/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-kube-api-access-vdrgl\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.752086 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.752180 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.752280 4725 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.754058 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-logs" (OuterVolumeSpecName: "logs") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.758053 4725 scope.go:117] "RemoveContainer" containerID="568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9" Dec 02 13:23:57 crc kubenswrapper[4725]: E1202 13:23:57.760944 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9\": container with ID starting with 568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9 not found: ID does not exist" containerID="568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.760989 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9"} err="failed to get container status \"568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9\": rpc error: code = NotFound desc = could not find container \"568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9\": container with ID starting with 568b6c810dfc69968560e0d1821d57aefad0f8569808cdb4f9ddcef91b82fec9 not found: ID does not exist" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.761024 4725 scope.go:117] "RemoveContainer" containerID="6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.762412 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-scripts" (OuterVolumeSpecName: "scripts") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.766123 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/153f44c9-3565-45bc-a162-8b7646df9cb2-kube-api-access-ccl4r" (OuterVolumeSpecName: "kube-api-access-ccl4r") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "kube-api-access-ccl4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: E1202 13:23:57.766276 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6\": container with ID starting with 6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6 not found: ID does not exist" containerID="6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.766311 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6"} err="failed to get container status \"6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6\": rpc error: code = NotFound desc = could not find container \"6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6\": container with ID starting with 6680323f8170283f4ca1b29a78e46734b0896a6c93dfc65b5a06fdab5869d5b6 not found: ID does not exist" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.766336 4725 scope.go:117] "RemoveContainer" containerID="6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.768251 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.794998 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.797948 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.798661 4725 scope.go:117] "RemoveContainer" containerID="e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.827797 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.840703 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-config-data" (OuterVolumeSpecName: "config-data") pod "153f44c9-3565-45bc-a162-8b7646df9cb2" (UID: "153f44c9-3565-45bc-a162-8b7646df9cb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855152 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855182 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855192 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855201 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855212 4725 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/153f44c9-3565-45bc-a162-8b7646df9cb2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855220 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/153f44c9-3565-45bc-a162-8b7646df9cb2-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855268 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.855279 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccl4r\" (UniqueName: \"kubernetes.io/projected/153f44c9-3565-45bc-a162-8b7646df9cb2-kube-api-access-ccl4r\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.862930 4725 scope.go:117] "RemoveContainer" containerID="e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.881540 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-config-data" (OuterVolumeSpecName: "config-data") pod "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" (UID: "06fe3c8a-ba63-42d9-ae09-3b62aadb7b21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.886106 4725 scope.go:117] "RemoveContainer" containerID="163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.920630 4725 scope.go:117] "RemoveContainer" containerID="6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f" Dec 02 13:23:57 crc kubenswrapper[4725]: E1202 13:23:57.921936 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f\": container with ID starting with 6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f not found: ID does not exist" containerID="6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.921980 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f"} err="failed to get container status \"6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f\": rpc error: code = NotFound desc = could not find container \"6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f\": container with ID starting with 6ba7a3bb481bbde32d6df9de997319de0c0444b4e64e3fa9bef650b911b3587f not found: ID does not exist" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.922063 4725 scope.go:117] "RemoveContainer" containerID="e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4" Dec 02 13:23:57 crc kubenswrapper[4725]: E1202 13:23:57.923300 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4\": container with ID starting with e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4 not found: ID does not exist" containerID="e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.923358 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4"} err="failed to get container status \"e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4\": rpc error: code = NotFound desc = could not find container \"e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4\": container with ID starting with e8d24973ebf123b0d6144f16267742cf29135855d483449b891d85efe9ca67d4 not found: ID does not exist" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.923391 4725 scope.go:117] "RemoveContainer" containerID="e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5" Dec 02 13:23:57 crc kubenswrapper[4725]: E1202 13:23:57.923739 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5\": container with ID starting with e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5 not found: ID does not exist" containerID="e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.923767 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5"} err="failed to get container status \"e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5\": rpc error: code = NotFound desc = could not find container \"e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5\": container with ID starting with e293359bdfd31909e9f1928b73866c34a0cbcf490c08911e787fe6832e562fd5 not found: ID does not exist" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.923802 4725 scope.go:117] "RemoveContainer" containerID="163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88" Dec 02 13:23:57 crc kubenswrapper[4725]: E1202 13:23:57.924146 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88\": container with ID starting with 163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88 not found: ID does not exist" containerID="163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.924190 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88"} err="failed to get container status \"163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88\": rpc error: code = NotFound desc = could not find container \"163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88\": container with ID starting with 163933c858733ff3a8e70f2d059978e0a63eb981eb52fe1b3b242ec9d4866b88 not found: ID does not exist" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.925321 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.957401 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.957436 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:23:57 crc kubenswrapper[4725]: W1202 13:23:57.961674 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b150790_de06_4a61_901f_014c70552a9c.slice/crio-cc327eb4874c9f482e62ad218f60aa88b32f6e1209e4d876a9f208235bbf484e WatchSource:0}: Error finding container cc327eb4874c9f482e62ad218f60aa88b32f6e1209e4d876a9f208235bbf484e: Status 404 returned error can't find the container with id cc327eb4874c9f482e62ad218f60aa88b32f6e1209e4d876a9f208235bbf484e Dec 02 13:23:57 crc kubenswrapper[4725]: I1202 13:23:57.965822 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.040729 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.061880 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.072886 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: E1202 13:23:58.073633 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-log" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.073721 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-log" Dec 02 13:23:58 crc kubenswrapper[4725]: E1202 13:23:58.073787 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-notification-agent" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.073859 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-notification-agent" Dec 02 13:23:58 crc kubenswrapper[4725]: E1202 13:23:58.073916 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-httpd" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.073969 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-httpd" Dec 02 13:23:58 crc kubenswrapper[4725]: E1202 13:23:58.074076 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="sg-core" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.074149 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="sg-core" Dec 02 13:23:58 crc kubenswrapper[4725]: E1202 13:23:58.074221 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="proxy-httpd" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.074288 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="proxy-httpd" Dec 02 13:23:58 crc kubenswrapper[4725]: E1202 13:23:58.074389 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-central-agent" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.074548 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-central-agent" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.074831 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-central-agent" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.074929 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="sg-core" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.075001 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-httpd" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.075079 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="proxy-httpd" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.075160 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" containerName="glance-log" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.075253 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" containerName="ceilometer-notification-agent" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.076648 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.082343 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.082359 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.104523 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.149607 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.161426 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.170263 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.172440 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.172661 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.172978 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.173063 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.173160 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvlkg\" (UniqueName: \"kubernetes.io/projected/17c38050-4a8a-45e9-8ff8-a40dffaee528-kube-api-access-rvlkg\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.173289 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17c38050-4a8a-45e9-8ff8-a40dffaee528-logs\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.173474 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17c38050-4a8a-45e9-8ff8-a40dffaee528-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.173604 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.173704 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.176559 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.176685 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.188117 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277597 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-run-httpd\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277668 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17c38050-4a8a-45e9-8ff8-a40dffaee528-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277705 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277724 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-scripts\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277749 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8tl5\" (UniqueName: \"kubernetes.io/projected/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-kube-api-access-t8tl5\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277775 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277793 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277877 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277905 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277925 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277954 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-config-data\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277970 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-log-httpd\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.277991 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvlkg\" (UniqueName: \"kubernetes.io/projected/17c38050-4a8a-45e9-8ff8-a40dffaee528-kube-api-access-rvlkg\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.278028 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17c38050-4a8a-45e9-8ff8-a40dffaee528-logs\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.278046 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.278518 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.279013 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17c38050-4a8a-45e9-8ff8-a40dffaee528-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.279046 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17c38050-4a8a-45e9-8ff8-a40dffaee528-logs\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.283020 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.284573 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.293806 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.295830 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvlkg\" (UniqueName: \"kubernetes.io/projected/17c38050-4a8a-45e9-8ff8-a40dffaee528-kube-api-access-rvlkg\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.298657 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c38050-4a8a-45e9-8ff8-a40dffaee528-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.307685 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"17c38050-4a8a-45e9-8ff8-a40dffaee528\") " pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379300 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-config-data\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379354 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-log-httpd\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379504 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379537 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-run-httpd\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379630 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-scripts\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379661 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8tl5\" (UniqueName: \"kubernetes.io/projected/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-kube-api-access-t8tl5\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.379692 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.381062 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-run-httpd\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.381186 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-log-httpd\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.383316 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.384696 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-config-data\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.385453 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-scripts\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.387671 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.407966 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8tl5\" (UniqueName: \"kubernetes.io/projected/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-kube-api-access-t8tl5\") pod \"ceilometer-0\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.522413 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.535589 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.740207 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b150790-de06-4a61-901f-014c70552a9c","Type":"ContainerStarted","Data":"6d50d7766737a8d9fac4501e602af6fb04ca79f14781768a8954e11c791b9c1e"} Dec 02 13:23:58 crc kubenswrapper[4725]: I1202 13:23:58.740540 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b150790-de06-4a61-901f-014c70552a9c","Type":"ContainerStarted","Data":"cc327eb4874c9f482e62ad218f60aa88b32f6e1209e4d876a9f208235bbf484e"} Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.196978 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.208258 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.299799 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06fe3c8a-ba63-42d9-ae09-3b62aadb7b21" path="/var/lib/kubelet/pods/06fe3c8a-ba63-42d9-ae09-3b62aadb7b21/volumes" Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.315715 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="153f44c9-3565-45bc-a162-8b7646df9cb2" path="/var/lib/kubelet/pods/153f44c9-3565-45bc-a162-8b7646df9cb2/volumes" Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.757454 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerStarted","Data":"40511ea6fbc12ca5643d241a3d3d191faf3d73b753b64d358f710ecc468a330c"} Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.759568 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17c38050-4a8a-45e9-8ff8-a40dffaee528","Type":"ContainerStarted","Data":"052170577b619793d46fdb04c0a7fae65d213a4f861a68fdd812de523ed0012f"} Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.764490 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b150790-de06-4a61-901f-014c70552a9c","Type":"ContainerStarted","Data":"d13c259ab5c01e633fa1d133e80049837d75b9c276ec5f5326c7ec566f876736"} Dec 02 13:23:59 crc kubenswrapper[4725]: I1202 13:23:59.790861 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.790827512 podStartE2EDuration="3.790827512s" podCreationTimestamp="2025-12-02 13:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:23:59.788173566 +0000 UTC m=+1170.744815271" watchObservedRunningTime="2025-12-02 13:23:59.790827512 +0000 UTC m=+1170.747469207" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.581516 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z6bgr"] Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.583347 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.587255 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.588109 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-zknx5" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.588569 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.599186 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z6bgr"] Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.633160 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.633297 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-config-data\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.633331 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-scripts\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.633354 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q59jr\" (UniqueName: \"kubernetes.io/projected/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-kube-api-access-q59jr\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.734541 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-config-data\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.734590 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-scripts\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.734871 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q59jr\" (UniqueName: \"kubernetes.io/projected/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-kube-api-access-q59jr\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.734954 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.751044 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-scripts\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.752418 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q59jr\" (UniqueName: \"kubernetes.io/projected/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-kube-api-access-q59jr\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.767614 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.768003 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-config-data\") pod \"nova-cell0-conductor-db-sync-z6bgr\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.774174 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerStarted","Data":"edaab83216fe37f375f276be0ec86f027d02ea5d3ddffef86bbdb64769597a84"} Dec 02 13:24:00 crc kubenswrapper[4725]: I1202 13:24:00.776667 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17c38050-4a8a-45e9-8ff8-a40dffaee528","Type":"ContainerStarted","Data":"09cefdf1b929b7a6f0966c7b2e57a4561207aa23cb3eb843205b4867c272b527"} Dec 02 13:24:01 crc kubenswrapper[4725]: I1202 13:24:01.037824 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:01 crc kubenswrapper[4725]: I1202 13:24:01.521535 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z6bgr"] Dec 02 13:24:01 crc kubenswrapper[4725]: W1202 13:24:01.523752 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2796bcd6_70d6_4895_9a2b_9d0de21ffa5e.slice/crio-439c5215cc33f85ea9f3235070e1be07a88c5aadca886bf2746d0a62b3892f73 WatchSource:0}: Error finding container 439c5215cc33f85ea9f3235070e1be07a88c5aadca886bf2746d0a62b3892f73: Status 404 returned error can't find the container with id 439c5215cc33f85ea9f3235070e1be07a88c5aadca886bf2746d0a62b3892f73 Dec 02 13:24:01 crc kubenswrapper[4725]: I1202 13:24:01.785832 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17c38050-4a8a-45e9-8ff8-a40dffaee528","Type":"ContainerStarted","Data":"a0328927e810240046c4b157dec854ee447ab54a05d4b6b44b67308b57f0fa51"} Dec 02 13:24:01 crc kubenswrapper[4725]: I1202 13:24:01.786876 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" event={"ID":"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e","Type":"ContainerStarted","Data":"439c5215cc33f85ea9f3235070e1be07a88c5aadca886bf2746d0a62b3892f73"} Dec 02 13:24:01 crc kubenswrapper[4725]: I1202 13:24:01.788654 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerStarted","Data":"64b16fbf9814de9b2ecb8a7511853d0a07798dfbdd1cfdf388a69bbb333d00b6"} Dec 02 13:24:01 crc kubenswrapper[4725]: I1202 13:24:01.806687 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.806667446 podStartE2EDuration="3.806667446s" podCreationTimestamp="2025-12-02 13:23:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:01.801077537 +0000 UTC m=+1172.757719232" watchObservedRunningTime="2025-12-02 13:24:01.806667446 +0000 UTC m=+1172.763309141" Dec 02 13:24:02 crc kubenswrapper[4725]: I1202 13:24:02.809613 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerStarted","Data":"49276e8ac61b5372db41f7d857a9bec63be42b3607764549a6a71bfa4cf1bb33"} Dec 02 13:24:04 crc kubenswrapper[4725]: I1202 13:24:04.842097 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerStarted","Data":"750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b"} Dec 02 13:24:04 crc kubenswrapper[4725]: I1202 13:24:04.844369 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:24:04 crc kubenswrapper[4725]: I1202 13:24:04.881554 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.536893933 podStartE2EDuration="6.881532151s" podCreationTimestamp="2025-12-02 13:23:58 +0000 UTC" firstStartedPulling="2025-12-02 13:23:59.218574603 +0000 UTC m=+1170.175216298" lastFinishedPulling="2025-12-02 13:24:03.563212821 +0000 UTC m=+1174.519854516" observedRunningTime="2025-12-02 13:24:04.864060943 +0000 UTC m=+1175.820702638" watchObservedRunningTime="2025-12-02 13:24:04.881532151 +0000 UTC m=+1175.838173846" Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.380073 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.380520 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.421776 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.436835 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.806900 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.867864 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-central-agent" containerID="cri-o://edaab83216fe37f375f276be0ec86f027d02ea5d3ddffef86bbdb64769597a84" gracePeriod=30 Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.868309 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-notification-agent" containerID="cri-o://64b16fbf9814de9b2ecb8a7511853d0a07798dfbdd1cfdf388a69bbb333d00b6" gracePeriod=30 Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.868394 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="sg-core" containerID="cri-o://49276e8ac61b5372db41f7d857a9bec63be42b3607764549a6a71bfa4cf1bb33" gracePeriod=30 Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.868321 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="proxy-httpd" containerID="cri-o://750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b" gracePeriod=30 Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.868541 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 13:24:07 crc kubenswrapper[4725]: I1202 13:24:07.868775 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 13:24:08 crc kubenswrapper[4725]: E1202 13:24:08.158385 4725 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod815fb1bf_e92e_46ce_8492_f1b3ab411b8c.slice/crio-conmon-750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod815fb1bf_e92e_46ce_8492_f1b3ab411b8c.slice/crio-750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b.scope\": RecentStats: unable to find data in memory cache]" Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.522777 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.522823 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.560052 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.587933 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879721 4725 generic.go:334] "Generic (PLEG): container finished" podID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerID="750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b" exitCode=0 Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879753 4725 generic.go:334] "Generic (PLEG): container finished" podID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerID="49276e8ac61b5372db41f7d857a9bec63be42b3607764549a6a71bfa4cf1bb33" exitCode=2 Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879763 4725 generic.go:334] "Generic (PLEG): container finished" podID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerID="64b16fbf9814de9b2ecb8a7511853d0a07798dfbdd1cfdf388a69bbb333d00b6" exitCode=0 Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879772 4725 generic.go:334] "Generic (PLEG): container finished" podID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerID="edaab83216fe37f375f276be0ec86f027d02ea5d3ddffef86bbdb64769597a84" exitCode=0 Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879789 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerDied","Data":"750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b"} Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879830 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerDied","Data":"49276e8ac61b5372db41f7d857a9bec63be42b3607764549a6a71bfa4cf1bb33"} Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879843 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerDied","Data":"64b16fbf9814de9b2ecb8a7511853d0a07798dfbdd1cfdf388a69bbb333d00b6"} Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.879854 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerDied","Data":"edaab83216fe37f375f276be0ec86f027d02ea5d3ddffef86bbdb64769597a84"} Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.880072 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:08 crc kubenswrapper[4725]: I1202 13:24:08.880095 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.130057 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.183396 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.183754 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228127 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-config-data\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228191 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-scripts\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228282 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-combined-ca-bundle\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228317 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-sg-core-conf-yaml\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228361 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8tl5\" (UniqueName: \"kubernetes.io/projected/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-kube-api-access-t8tl5\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228386 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-run-httpd\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.228534 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-log-httpd\") pod \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\" (UID: \"815fb1bf-e92e-46ce-8492-f1b3ab411b8c\") " Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.229474 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.230976 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.238085 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-kube-api-access-t8tl5" (OuterVolumeSpecName: "kube-api-access-t8tl5") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "kube-api-access-t8tl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.244535 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-scripts" (OuterVolumeSpecName: "scripts") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.271211 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.293760 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.313528 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.332376 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.332412 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.332429 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.332438 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.332448 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8tl5\" (UniqueName: \"kubernetes.io/projected/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-kube-api-access-t8tl5\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.332471 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.390698 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-config-data" (OuterVolumeSpecName: "config-data") pod "815fb1bf-e92e-46ce-8492-f1b3ab411b8c" (UID: "815fb1bf-e92e-46ce-8492-f1b3ab411b8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.434982 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815fb1bf-e92e-46ce-8492-f1b3ab411b8c-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.925895 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"815fb1bf-e92e-46ce-8492-f1b3ab411b8c","Type":"ContainerDied","Data":"40511ea6fbc12ca5643d241a3d3d191faf3d73b753b64d358f710ecc468a330c"} Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.925963 4725 scope.go:117] "RemoveContainer" containerID="750bdc7c54ce17b197a27bbfcebab369f8497a0efc10bbe5d8b3a4028f31a81b" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.926234 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.927586 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" event={"ID":"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e","Type":"ContainerStarted","Data":"942f19c9cb8648563e6cc8eec3980e070e4e79e53f79d2029cd65413b9468f41"} Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.957604 4725 scope.go:117] "RemoveContainer" containerID="49276e8ac61b5372db41f7d857a9bec63be42b3607764549a6a71bfa4cf1bb33" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.968118 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" podStartSLOduration=2.674563563 podStartE2EDuration="10.968080218s" podCreationTimestamp="2025-12-02 13:24:00 +0000 UTC" firstStartedPulling="2025-12-02 13:24:01.52588 +0000 UTC m=+1172.482521705" lastFinishedPulling="2025-12-02 13:24:09.819396665 +0000 UTC m=+1180.776038360" observedRunningTime="2025-12-02 13:24:10.947980385 +0000 UTC m=+1181.904622080" watchObservedRunningTime="2025-12-02 13:24:10.968080218 +0000 UTC m=+1181.924721903" Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.984880 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:10 crc kubenswrapper[4725]: I1202 13:24:10.988675 4725 scope.go:117] "RemoveContainer" containerID="64b16fbf9814de9b2ecb8a7511853d0a07798dfbdd1cfdf388a69bbb333d00b6" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.010174 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.015642 4725 scope.go:117] "RemoveContainer" containerID="edaab83216fe37f375f276be0ec86f027d02ea5d3ddffef86bbdb64769597a84" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.024517 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:11 crc kubenswrapper[4725]: E1202 13:24:11.024853 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="sg-core" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.024864 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="sg-core" Dec 02 13:24:11 crc kubenswrapper[4725]: E1202 13:24:11.024881 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="proxy-httpd" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.024887 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="proxy-httpd" Dec 02 13:24:11 crc kubenswrapper[4725]: E1202 13:24:11.024903 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-notification-agent" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.024909 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-notification-agent" Dec 02 13:24:11 crc kubenswrapper[4725]: E1202 13:24:11.024925 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-central-agent" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.024930 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-central-agent" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.025093 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-central-agent" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.025107 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="ceilometer-notification-agent" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.025119 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="proxy-httpd" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.025140 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" containerName="sg-core" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.026784 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.029689 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.029859 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.032948 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149070 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149154 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-run-httpd\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149212 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-log-httpd\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149286 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149375 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-config-data\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149474 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-scripts\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.149527 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7c28\" (UniqueName: \"kubernetes.io/projected/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-kube-api-access-m7c28\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.247386 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.247563 4725 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251070 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-config-data\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251136 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-scripts\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251177 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7c28\" (UniqueName: \"kubernetes.io/projected/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-kube-api-access-m7c28\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251213 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251252 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-run-httpd\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251282 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-log-httpd\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251318 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251825 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-log-httpd\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.251898 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-run-httpd\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.256083 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.256777 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-scripts\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.260373 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-config-data\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.270025 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.278691 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="815fb1bf-e92e-46ce-8492-f1b3ab411b8c" path="/var/lib/kubelet/pods/815fb1bf-e92e-46ce-8492-f1b3ab411b8c/volumes" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.281171 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7c28\" (UniqueName: \"kubernetes.io/projected/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-kube-api-access-m7c28\") pod \"ceilometer-0\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.285368 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.371126 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.925753 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:11 crc kubenswrapper[4725]: I1202 13:24:11.945709 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerStarted","Data":"982cdb17a7d9bae4fc46057b8c0d934cd8a31656bbd37e3a9599bdb6cb1f0c09"} Dec 02 13:24:12 crc kubenswrapper[4725]: I1202 13:24:12.965395 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerStarted","Data":"0ac6b72baf26b034e2331ed06d4e6f8dc65fd872083df74a82bc77772f07f482"} Dec 02 13:24:13 crc kubenswrapper[4725]: I1202 13:24:13.977148 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerStarted","Data":"5381c380d7ff1a0de5c1da58b660a78160743027d1ff9ce611c0682fbe914dd3"} Dec 02 13:24:14 crc kubenswrapper[4725]: I1202 13:24:14.991387 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerStarted","Data":"bd5b83b7bb70d4c58e86a17406f4109f74050af8c069b943d87ddec9083c52da"} Dec 02 13:24:17 crc kubenswrapper[4725]: I1202 13:24:17.026006 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerStarted","Data":"62e4b0bbe0a93f4db166ad5185700e98ab470446aa586e83b5af44ca3f792b85"} Dec 02 13:24:17 crc kubenswrapper[4725]: I1202 13:24:17.026442 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:24:17 crc kubenswrapper[4725]: I1202 13:24:17.051371 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.748494551 podStartE2EDuration="7.051326034s" podCreationTimestamp="2025-12-02 13:24:10 +0000 UTC" firstStartedPulling="2025-12-02 13:24:11.933255841 +0000 UTC m=+1182.889897536" lastFinishedPulling="2025-12-02 13:24:16.236087324 +0000 UTC m=+1187.192729019" observedRunningTime="2025-12-02 13:24:17.042837942 +0000 UTC m=+1187.999479647" watchObservedRunningTime="2025-12-02 13:24:17.051326034 +0000 UTC m=+1188.007967729" Dec 02 13:24:25 crc kubenswrapper[4725]: I1202 13:24:25.115372 4725 generic.go:334] "Generic (PLEG): container finished" podID="2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" containerID="942f19c9cb8648563e6cc8eec3980e070e4e79e53f79d2029cd65413b9468f41" exitCode=0 Dec 02 13:24:25 crc kubenswrapper[4725]: I1202 13:24:25.115474 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" event={"ID":"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e","Type":"ContainerDied","Data":"942f19c9cb8648563e6cc8eec3980e070e4e79e53f79d2029cd65413b9468f41"} Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.435027 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.471283 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-combined-ca-bundle\") pod \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.471406 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q59jr\" (UniqueName: \"kubernetes.io/projected/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-kube-api-access-q59jr\") pod \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.471523 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-config-data\") pod \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.471571 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-scripts\") pod \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\" (UID: \"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e\") " Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.478192 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-kube-api-access-q59jr" (OuterVolumeSpecName: "kube-api-access-q59jr") pod "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" (UID: "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e"). InnerVolumeSpecName "kube-api-access-q59jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.480831 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-scripts" (OuterVolumeSpecName: "scripts") pod "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" (UID: "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.504994 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" (UID: "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.506652 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-config-data" (OuterVolumeSpecName: "config-data") pod "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" (UID: "2796bcd6-70d6-4895-9a2b-9d0de21ffa5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.573798 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.573995 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.574082 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:26 crc kubenswrapper[4725]: I1202 13:24:26.574153 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q59jr\" (UniqueName: \"kubernetes.io/projected/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e-kube-api-access-q59jr\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.133138 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" event={"ID":"2796bcd6-70d6-4895-9a2b-9d0de21ffa5e","Type":"ContainerDied","Data":"439c5215cc33f85ea9f3235070e1be07a88c5aadca886bf2746d0a62b3892f73"} Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.133181 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="439c5215cc33f85ea9f3235070e1be07a88c5aadca886bf2746d0a62b3892f73" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.133245 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z6bgr" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.232127 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 13:24:27 crc kubenswrapper[4725]: E1202 13:24:27.233043 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" containerName="nova-cell0-conductor-db-sync" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.233065 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" containerName="nova-cell0-conductor-db-sync" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.233278 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" containerName="nova-cell0-conductor-db-sync" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.233948 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.239732 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.239884 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-zknx5" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.251641 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.290451 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.290566 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbxnn\" (UniqueName: \"kubernetes.io/projected/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-kube-api-access-fbxnn\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.290597 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.392402 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbxnn\" (UniqueName: \"kubernetes.io/projected/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-kube-api-access-fbxnn\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.392797 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.392961 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.398810 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.402052 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.411267 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbxnn\" (UniqueName: \"kubernetes.io/projected/84138030-0cbd-41bd-8fb4-afe9fcf57bbc-kube-api-access-fbxnn\") pod \"nova-cell0-conductor-0\" (UID: \"84138030-0cbd-41bd-8fb4-afe9fcf57bbc\") " pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:27 crc kubenswrapper[4725]: I1202 13:24:27.561234 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:28 crc kubenswrapper[4725]: I1202 13:24:28.002304 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 13:24:28 crc kubenswrapper[4725]: I1202 13:24:28.146158 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"84138030-0cbd-41bd-8fb4-afe9fcf57bbc","Type":"ContainerStarted","Data":"32253aea30e90b10cdc6bced7926a4780753b511b099a1ac57ef5df64272308a"} Dec 02 13:24:29 crc kubenswrapper[4725]: I1202 13:24:29.154829 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"84138030-0cbd-41bd-8fb4-afe9fcf57bbc","Type":"ContainerStarted","Data":"b53de2ab9e73441b0ce60b3135fd4dcde730c3ed813bba2b1898561f29030c14"} Dec 02 13:24:29 crc kubenswrapper[4725]: I1202 13:24:29.155313 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:29 crc kubenswrapper[4725]: I1202 13:24:29.172434 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.172419228 podStartE2EDuration="2.172419228s" podCreationTimestamp="2025-12-02 13:24:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:29.169339791 +0000 UTC m=+1200.125981486" watchObservedRunningTime="2025-12-02 13:24:29.172419228 +0000 UTC m=+1200.129060923" Dec 02 13:24:37 crc kubenswrapper[4725]: I1202 13:24:37.589451 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.078717 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-cltvz"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.089725 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-cltvz"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.080139 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.096678 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.096988 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.231971 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgftn\" (UniqueName: \"kubernetes.io/projected/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-kube-api-access-dgftn\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.232065 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-scripts\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.232653 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-config-data\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.232679 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.296582 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.301953 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.308906 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.310364 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.314898 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.315610 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.333976 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-config-data\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.334039 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.334091 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgftn\" (UniqueName: \"kubernetes.io/projected/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-kube-api-access-dgftn\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.334184 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-scripts\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.343621 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.346844 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.381135 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-config-data\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.382192 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-scripts\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.382795 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgftn\" (UniqueName: \"kubernetes.io/projected/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-kube-api-access-dgftn\") pod \"nova-cell0-cell-mapping-cltvz\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.390634 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.392257 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.419886 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.428336 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436578 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9xbz\" (UniqueName: \"kubernetes.io/projected/2c05245b-3599-4732-a65e-cbfda339ca53-kube-api-access-c9xbz\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436626 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436657 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c05245b-3599-4732-a65e-cbfda339ca53-logs\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436678 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-config-data\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436726 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436789 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-config-data\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.436805 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjdkg\" (UniqueName: \"kubernetes.io/projected/cd7fc68c-2871-4b3f-96bd-50d9274853e5-kube-api-access-pjdkg\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.452108 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.529752 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547548 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-logs\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547641 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-config-data\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547675 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9xbz\" (UniqueName: \"kubernetes.io/projected/2c05245b-3599-4732-a65e-cbfda339ca53-kube-api-access-c9xbz\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547709 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547744 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c05245b-3599-4732-a65e-cbfda339ca53-logs\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547765 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-config-data\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547790 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547918 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxd9t\" (UniqueName: \"kubernetes.io/projected/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-kube-api-access-jxd9t\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547948 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-config-data\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547963 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.547991 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjdkg\" (UniqueName: \"kubernetes.io/projected/cd7fc68c-2871-4b3f-96bd-50d9274853e5-kube-api-access-pjdkg\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.550816 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c05245b-3599-4732-a65e-cbfda339ca53-logs\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.570615 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.573290 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-config-data\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.574101 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-config-data\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.588379 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.616599 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.618090 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.644232 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9xbz\" (UniqueName: \"kubernetes.io/projected/2c05245b-3599-4732-a65e-cbfda339ca53-kube-api-access-c9xbz\") pod \"nova-api-0\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.647182 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.651687 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-logs\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.661115 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-config-data\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.661598 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxd9t\" (UniqueName: \"kubernetes.io/projected/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-kube-api-access-jxd9t\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.661757 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.654403 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-logs\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.665093 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjdkg\" (UniqueName: \"kubernetes.io/projected/cd7fc68c-2871-4b3f-96bd-50d9274853e5-kube-api-access-pjdkg\") pod \"nova-scheduler-0\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.665161 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gb8s7"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.667601 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.684239 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.701344 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-config-data\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.727965 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.736442 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxd9t\" (UniqueName: \"kubernetes.io/projected/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-kube-api-access-jxd9t\") pod \"nova-metadata-0\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " pod="openstack/nova-metadata-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.761222 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gb8s7"] Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764359 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764494 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764576 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-config\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764612 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764698 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764730 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kxwb\" (UniqueName: \"kubernetes.io/projected/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-kube-api-access-5kxwb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764763 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764805 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.764846 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxz7c\" (UniqueName: \"kubernetes.io/projected/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-kube-api-access-mxz7c\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.857879 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866061 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866134 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866186 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-config\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866224 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866523 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866559 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kxwb\" (UniqueName: \"kubernetes.io/projected/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-kube-api-access-5kxwb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866585 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866615 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.866645 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxz7c\" (UniqueName: \"kubernetes.io/projected/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-kube-api-access-mxz7c\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.869962 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.870001 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.871085 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.871099 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.872162 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.873957 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.876447 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-config\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.883420 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxz7c\" (UniqueName: \"kubernetes.io/projected/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-kube-api-access-mxz7c\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.894698 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kxwb\" (UniqueName: \"kubernetes.io/projected/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-kube-api-access-5kxwb\") pod \"dnsmasq-dns-757b4f8459-gb8s7\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.945394 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:24:38 crc kubenswrapper[4725]: I1202 13:24:38.967394 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.051908 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.075878 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.162212 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-cltvz"] Dec 02 13:24:39 crc kubenswrapper[4725]: W1202 13:24:39.234596 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f6eb9bb_6946_4475_9421_8ef60d3cb50b.slice/crio-abc3e201017f0cb80df1a48e135a94cee9546d9df744e405cb928ebf0fc811cd WatchSource:0}: Error finding container abc3e201017f0cb80df1a48e135a94cee9546d9df744e405cb928ebf0fc811cd: Status 404 returned error can't find the container with id abc3e201017f0cb80df1a48e135a94cee9546d9df744e405cb928ebf0fc811cd Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.332450 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cltvz" event={"ID":"6f6eb9bb-6946-4475-9421-8ef60d3cb50b","Type":"ContainerStarted","Data":"abc3e201017f0cb80df1a48e135a94cee9546d9df744e405cb928ebf0fc811cd"} Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.408617 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.427916 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.453036 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-sffb7"] Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.454445 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.468548 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-sffb7"] Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.469262 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.485869 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.518123 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhsqr\" (UniqueName: \"kubernetes.io/projected/a606a54a-2bb1-4390-8682-dd7e78b4e955-kube-api-access-hhsqr\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.518327 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-config-data\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.518366 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.518420 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-scripts\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: W1202 13:24:39.526951 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c05245b_3599_4732_a65e_cbfda339ca53.slice/crio-0a701eedba18918b4edab06f16d52f59c3161f3c552b514db2a5eef61af3906b WatchSource:0}: Error finding container 0a701eedba18918b4edab06f16d52f59c3161f3c552b514db2a5eef61af3906b: Status 404 returned error can't find the container with id 0a701eedba18918b4edab06f16d52f59c3161f3c552b514db2a5eef61af3906b Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.543188 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.620599 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhsqr\" (UniqueName: \"kubernetes.io/projected/a606a54a-2bb1-4390-8682-dd7e78b4e955-kube-api-access-hhsqr\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.621310 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-config-data\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.621373 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.621553 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-scripts\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.630616 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-scripts\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.631380 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-config-data\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.632486 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.660920 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhsqr\" (UniqueName: \"kubernetes.io/projected/a606a54a-2bb1-4390-8682-dd7e78b4e955-kube-api-access-hhsqr\") pod \"nova-cell1-conductor-db-sync-sffb7\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.682638 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.733545 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.775414 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gb8s7"] Dec 02 13:24:39 crc kubenswrapper[4725]: W1202 13:24:39.782338 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c41b73e_5453_4e4f_8f35_dcb324b78e4e.slice/crio-443bb844e045c4ad66122ada9ac01197e4f31f88392968ff97dd59ee3b00ed9f WatchSource:0}: Error finding container 443bb844e045c4ad66122ada9ac01197e4f31f88392968ff97dd59ee3b00ed9f: Status 404 returned error can't find the container with id 443bb844e045c4ad66122ada9ac01197e4f31f88392968ff97dd59ee3b00ed9f Dec 02 13:24:39 crc kubenswrapper[4725]: I1202 13:24:39.784910 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.261658 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-sffb7"] Dec 02 13:24:40 crc kubenswrapper[4725]: W1202 13:24:40.272497 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda606a54a_2bb1_4390_8682_dd7e78b4e955.slice/crio-2bafcec17c298286035edb8a56bef66e6bd6149bf18e637a65dc563ff297ac67 WatchSource:0}: Error finding container 2bafcec17c298286035edb8a56bef66e6bd6149bf18e637a65dc563ff297ac67: Status 404 returned error can't find the container with id 2bafcec17c298286035edb8a56bef66e6bd6149bf18e637a65dc563ff297ac67 Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.329202 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd7fc68c-2871-4b3f-96bd-50d9274853e5","Type":"ContainerStarted","Data":"22730c89e0ec25818b29cf9db3ca6742f62999beb6c4e2f97dda0585d7103b0f"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.331357 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-sffb7" event={"ID":"a606a54a-2bb1-4390-8682-dd7e78b4e955","Type":"ContainerStarted","Data":"2bafcec17c298286035edb8a56bef66e6bd6149bf18e637a65dc563ff297ac67"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.333243 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d0475c71-2182-49ca-9bc0-cf0a4e278e8b","Type":"ContainerStarted","Data":"06333a02de5964c4fcf28a5028ed04d56f763a6a0d50e131fdb2cf790092ab21"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.340470 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cltvz" event={"ID":"6f6eb9bb-6946-4475-9421-8ef60d3cb50b","Type":"ContainerStarted","Data":"2f7f59366c6df9411bd2995749cd6abbc1d98a0170ea697ab1955c05a1b0f28e"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.345127 4725 generic.go:334] "Generic (PLEG): container finished" podID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerID="09e8b28de4ba97b70df54d62e87722401e110d7e731fd53e3379c3e0cb14cbfc" exitCode=0 Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.345194 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" event={"ID":"7c41b73e-5453-4e4f-8f35-dcb324b78e4e","Type":"ContainerDied","Data":"09e8b28de4ba97b70df54d62e87722401e110d7e731fd53e3379c3e0cb14cbfc"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.345221 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" event={"ID":"7c41b73e-5453-4e4f-8f35-dcb324b78e4e","Type":"ContainerStarted","Data":"443bb844e045c4ad66122ada9ac01197e4f31f88392968ff97dd59ee3b00ed9f"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.350384 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28","Type":"ContainerStarted","Data":"ea5321566e0b5282acba089f46366746ba3aff6eeb43dd0df3634339a8847c20"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.364213 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c05245b-3599-4732-a65e-cbfda339ca53","Type":"ContainerStarted","Data":"0a701eedba18918b4edab06f16d52f59c3161f3c552b514db2a5eef61af3906b"} Dec 02 13:24:40 crc kubenswrapper[4725]: I1202 13:24:40.379262 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-cltvz" podStartSLOduration=2.379230213 podStartE2EDuration="2.379230213s" podCreationTimestamp="2025-12-02 13:24:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:40.362274538 +0000 UTC m=+1211.318916233" watchObservedRunningTime="2025-12-02 13:24:40.379230213 +0000 UTC m=+1211.335871908" Dec 02 13:24:41 crc kubenswrapper[4725]: I1202 13:24:41.375821 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 02 13:24:41 crc kubenswrapper[4725]: I1202 13:24:41.382001 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-sffb7" event={"ID":"a606a54a-2bb1-4390-8682-dd7e78b4e955","Type":"ContainerStarted","Data":"db5cff9dcd14e1d46ebb9ccec66be80fb821bdc98fcb6c2d1debf9cff2c21a17"} Dec 02 13:24:41 crc kubenswrapper[4725]: I1202 13:24:41.415241 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-sffb7" podStartSLOduration=2.415221047 podStartE2EDuration="2.415221047s" podCreationTimestamp="2025-12-02 13:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:41.414858918 +0000 UTC m=+1212.371500613" watchObservedRunningTime="2025-12-02 13:24:41.415221047 +0000 UTC m=+1212.371862742" Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.414183 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d0475c71-2182-49ca-9bc0-cf0a4e278e8b","Type":"ContainerStarted","Data":"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671"} Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.421168 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" event={"ID":"7c41b73e-5453-4e4f-8f35-dcb324b78e4e","Type":"ContainerStarted","Data":"66c3b969706fc07653030d20d28aeb6ab12b72fbfdd14d4081ecc667d8220507"} Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.422428 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.429079 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c05245b-3599-4732-a65e-cbfda339ca53","Type":"ContainerStarted","Data":"42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330"} Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.435591 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd7fc68c-2871-4b3f-96bd-50d9274853e5","Type":"ContainerStarted","Data":"2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c"} Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.445515 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" podStartSLOduration=4.445497038 podStartE2EDuration="4.445497038s" podCreationTimestamp="2025-12-02 13:24:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:42.443507038 +0000 UTC m=+1213.400148733" watchObservedRunningTime="2025-12-02 13:24:42.445497038 +0000 UTC m=+1213.402138733" Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.470380 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.064189028 podStartE2EDuration="4.47036134s" podCreationTimestamp="2025-12-02 13:24:38 +0000 UTC" firstStartedPulling="2025-12-02 13:24:39.42764368 +0000 UTC m=+1210.384285365" lastFinishedPulling="2025-12-02 13:24:41.833815982 +0000 UTC m=+1212.790457677" observedRunningTime="2025-12-02 13:24:42.462089203 +0000 UTC m=+1213.418730898" watchObservedRunningTime="2025-12-02 13:24:42.47036134 +0000 UTC m=+1213.427003035" Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.760400 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:42 crc kubenswrapper[4725]: I1202 13:24:42.785574 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.452993 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c05245b-3599-4732-a65e-cbfda339ca53","Type":"ContainerStarted","Data":"540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23"} Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.458993 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d0475c71-2182-49ca-9bc0-cf0a4e278e8b","Type":"ContainerStarted","Data":"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8"} Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.459314 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-log" containerID="cri-o://482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671" gracePeriod=30 Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.459320 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-metadata" containerID="cri-o://10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8" gracePeriod=30 Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.480884 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.17159951 podStartE2EDuration="5.480865707s" podCreationTimestamp="2025-12-02 13:24:38 +0000 UTC" firstStartedPulling="2025-12-02 13:24:39.533995892 +0000 UTC m=+1210.490637587" lastFinishedPulling="2025-12-02 13:24:41.843262089 +0000 UTC m=+1212.799903784" observedRunningTime="2025-12-02 13:24:43.478931058 +0000 UTC m=+1214.435572753" watchObservedRunningTime="2025-12-02 13:24:43.480865707 +0000 UTC m=+1214.437507422" Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.527307 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.383315028 podStartE2EDuration="5.527280748s" podCreationTimestamp="2025-12-02 13:24:38 +0000 UTC" firstStartedPulling="2025-12-02 13:24:39.690762875 +0000 UTC m=+1210.647404570" lastFinishedPulling="2025-12-02 13:24:41.834728595 +0000 UTC m=+1212.791370290" observedRunningTime="2025-12-02 13:24:43.504118949 +0000 UTC m=+1214.460760664" watchObservedRunningTime="2025-12-02 13:24:43.527280748 +0000 UTC m=+1214.483922443" Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.858171 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.967887 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:24:43 crc kubenswrapper[4725]: I1202 13:24:43.968320 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.090811 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.123367 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-combined-ca-bundle\") pod \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.123530 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-logs\") pod \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.123671 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-config-data\") pod \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.123780 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxd9t\" (UniqueName: \"kubernetes.io/projected/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-kube-api-access-jxd9t\") pod \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\" (UID: \"d0475c71-2182-49ca-9bc0-cf0a4e278e8b\") " Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.124231 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-logs" (OuterVolumeSpecName: "logs") pod "d0475c71-2182-49ca-9bc0-cf0a4e278e8b" (UID: "d0475c71-2182-49ca-9bc0-cf0a4e278e8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.130394 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-kube-api-access-jxd9t" (OuterVolumeSpecName: "kube-api-access-jxd9t") pod "d0475c71-2182-49ca-9bc0-cf0a4e278e8b" (UID: "d0475c71-2182-49ca-9bc0-cf0a4e278e8b"). InnerVolumeSpecName "kube-api-access-jxd9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.154285 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-config-data" (OuterVolumeSpecName: "config-data") pod "d0475c71-2182-49ca-9bc0-cf0a4e278e8b" (UID: "d0475c71-2182-49ca-9bc0-cf0a4e278e8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.159381 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0475c71-2182-49ca-9bc0-cf0a4e278e8b" (UID: "d0475c71-2182-49ca-9bc0-cf0a4e278e8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.226289 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxd9t\" (UniqueName: \"kubernetes.io/projected/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-kube-api-access-jxd9t\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.226329 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.226341 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.226355 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0475c71-2182-49ca-9bc0-cf0a4e278e8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.469473 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28","Type":"ContainerStarted","Data":"78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757"} Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.469561 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757" gracePeriod=30 Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474500 4725 generic.go:334] "Generic (PLEG): container finished" podID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerID="10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8" exitCode=0 Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474528 4725 generic.go:334] "Generic (PLEG): container finished" podID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerID="482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671" exitCode=143 Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474550 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d0475c71-2182-49ca-9bc0-cf0a4e278e8b","Type":"ContainerDied","Data":"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8"} Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474584 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d0475c71-2182-49ca-9bc0-cf0a4e278e8b","Type":"ContainerDied","Data":"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671"} Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474594 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d0475c71-2182-49ca-9bc0-cf0a4e278e8b","Type":"ContainerDied","Data":"06333a02de5964c4fcf28a5028ed04d56f763a6a0d50e131fdb2cf790092ab21"} Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474612 4725 scope.go:117] "RemoveContainer" containerID="10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.474640 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.496745 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.878163248 podStartE2EDuration="6.496728528s" podCreationTimestamp="2025-12-02 13:24:38 +0000 UTC" firstStartedPulling="2025-12-02 13:24:39.809106726 +0000 UTC m=+1210.765748421" lastFinishedPulling="2025-12-02 13:24:43.427672006 +0000 UTC m=+1214.384313701" observedRunningTime="2025-12-02 13:24:44.491445806 +0000 UTC m=+1215.448087511" watchObservedRunningTime="2025-12-02 13:24:44.496728528 +0000 UTC m=+1215.453370223" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.508620 4725 scope.go:117] "RemoveContainer" containerID="482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.530107 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.535193 4725 scope.go:117] "RemoveContainer" containerID="10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8" Dec 02 13:24:44 crc kubenswrapper[4725]: E1202 13:24:44.535727 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8\": container with ID starting with 10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8 not found: ID does not exist" containerID="10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.535769 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8"} err="failed to get container status \"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8\": rpc error: code = NotFound desc = could not find container \"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8\": container with ID starting with 10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8 not found: ID does not exist" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.535797 4725 scope.go:117] "RemoveContainer" containerID="482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671" Dec 02 13:24:44 crc kubenswrapper[4725]: E1202 13:24:44.536039 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671\": container with ID starting with 482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671 not found: ID does not exist" containerID="482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.536070 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671"} err="failed to get container status \"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671\": rpc error: code = NotFound desc = could not find container \"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671\": container with ID starting with 482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671 not found: ID does not exist" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.536083 4725 scope.go:117] "RemoveContainer" containerID="10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.536313 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8"} err="failed to get container status \"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8\": rpc error: code = NotFound desc = could not find container \"10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8\": container with ID starting with 10985d425c3a10080172d0ba767773e18e87aebdfb28810c54b2465de1fbbea8 not found: ID does not exist" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.536336 4725 scope.go:117] "RemoveContainer" containerID="482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.539853 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671"} err="failed to get container status \"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671\": rpc error: code = NotFound desc = could not find container \"482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671\": container with ID starting with 482c51145c8d238e203af59d34b981618085ec9bda9c8eabd6068c361b147671 not found: ID does not exist" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.544700 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.562619 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:44 crc kubenswrapper[4725]: E1202 13:24:44.563152 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-metadata" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.563174 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-metadata" Dec 02 13:24:44 crc kubenswrapper[4725]: E1202 13:24:44.563196 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-log" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.563206 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-log" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.563446 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-log" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.563492 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" containerName="nova-metadata-metadata" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.564795 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.568326 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.568572 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.580909 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.635996 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-config-data\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.637263 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.637409 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.637670 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1f227c-8fae-46fc-9a85-921c838c9af3-logs\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.637737 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn5zz\" (UniqueName: \"kubernetes.io/projected/ec1f227c-8fae-46fc-9a85-921c838c9af3-kube-api-access-fn5zz\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.739712 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-config-data\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.739804 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.739842 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.739902 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1f227c-8fae-46fc-9a85-921c838c9af3-logs\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.739924 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn5zz\" (UniqueName: \"kubernetes.io/projected/ec1f227c-8fae-46fc-9a85-921c838c9af3-kube-api-access-fn5zz\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.740648 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1f227c-8fae-46fc-9a85-921c838c9af3-logs\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.746818 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.749091 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.756397 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-config-data\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.772202 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn5zz\" (UniqueName: \"kubernetes.io/projected/ec1f227c-8fae-46fc-9a85-921c838c9af3-kube-api-access-fn5zz\") pod \"nova-metadata-0\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " pod="openstack/nova-metadata-0" Dec 02 13:24:44 crc kubenswrapper[4725]: I1202 13:24:44.887312 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:45 crc kubenswrapper[4725]: I1202 13:24:45.279656 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0475c71-2182-49ca-9bc0-cf0a4e278e8b" path="/var/lib/kubelet/pods/d0475c71-2182-49ca-9bc0-cf0a4e278e8b/volumes" Dec 02 13:24:45 crc kubenswrapper[4725]: I1202 13:24:45.424131 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:45 crc kubenswrapper[4725]: I1202 13:24:45.493426 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1f227c-8fae-46fc-9a85-921c838c9af3","Type":"ContainerStarted","Data":"4ed8c257032687be7f0f5557a08a318adebad6841710ec3cde9d940b4c1708e2"} Dec 02 13:24:46 crc kubenswrapper[4725]: I1202 13:24:46.425321 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:24:46 crc kubenswrapper[4725]: I1202 13:24:46.425740 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1bbe1936-05fb-4f72-affd-7573be426c6e" containerName="kube-state-metrics" containerID="cri-o://a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1" gracePeriod=30 Dec 02 13:24:46 crc kubenswrapper[4725]: I1202 13:24:46.512984 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1f227c-8fae-46fc-9a85-921c838c9af3","Type":"ContainerStarted","Data":"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a"} Dec 02 13:24:46 crc kubenswrapper[4725]: I1202 13:24:46.513037 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1f227c-8fae-46fc-9a85-921c838c9af3","Type":"ContainerStarted","Data":"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0"} Dec 02 13:24:46 crc kubenswrapper[4725]: I1202 13:24:46.537243 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.537215758 podStartE2EDuration="2.537215758s" podCreationTimestamp="2025-12-02 13:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:46.531275949 +0000 UTC m=+1217.487917644" watchObservedRunningTime="2025-12-02 13:24:46.537215758 +0000 UTC m=+1217.493857473" Dec 02 13:24:46 crc kubenswrapper[4725]: I1202 13:24:46.969012 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.005023 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfszl\" (UniqueName: \"kubernetes.io/projected/1bbe1936-05fb-4f72-affd-7573be426c6e-kube-api-access-gfszl\") pod \"1bbe1936-05fb-4f72-affd-7573be426c6e\" (UID: \"1bbe1936-05fb-4f72-affd-7573be426c6e\") " Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.013022 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bbe1936-05fb-4f72-affd-7573be426c6e-kube-api-access-gfszl" (OuterVolumeSpecName: "kube-api-access-gfszl") pod "1bbe1936-05fb-4f72-affd-7573be426c6e" (UID: "1bbe1936-05fb-4f72-affd-7573be426c6e"). InnerVolumeSpecName "kube-api-access-gfszl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.107946 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfszl\" (UniqueName: \"kubernetes.io/projected/1bbe1936-05fb-4f72-affd-7573be426c6e-kube-api-access-gfszl\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.525344 4725 generic.go:334] "Generic (PLEG): container finished" podID="1bbe1936-05fb-4f72-affd-7573be426c6e" containerID="a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1" exitCode=2 Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.526658 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.527104 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bbe1936-05fb-4f72-affd-7573be426c6e","Type":"ContainerDied","Data":"a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1"} Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.527132 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1bbe1936-05fb-4f72-affd-7573be426c6e","Type":"ContainerDied","Data":"6f042378bed374a085970210d8d4b7886ba945df28b7a2defd98e13928b6e590"} Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.527152 4725 scope.go:117] "RemoveContainer" containerID="a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.550654 4725 scope.go:117] "RemoveContainer" containerID="a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1" Dec 02 13:24:47 crc kubenswrapper[4725]: E1202 13:24:47.551342 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1\": container with ID starting with a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1 not found: ID does not exist" containerID="a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.551373 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1"} err="failed to get container status \"a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1\": rpc error: code = NotFound desc = could not find container \"a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1\": container with ID starting with a1f48328bb0a33b33561006b2aaee4f69547c78aceed60f925d2bd3d830cccc1 not found: ID does not exist" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.570759 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.585215 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.596026 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:24:47 crc kubenswrapper[4725]: E1202 13:24:47.596599 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bbe1936-05fb-4f72-affd-7573be426c6e" containerName="kube-state-metrics" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.596617 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bbe1936-05fb-4f72-affd-7573be426c6e" containerName="kube-state-metrics" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.596829 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bbe1936-05fb-4f72-affd-7573be426c6e" containerName="kube-state-metrics" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.597615 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.602995 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.603036 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.623521 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.636726 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.636781 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhk9\" (UniqueName: \"kubernetes.io/projected/a093384f-2994-4207-8447-1cfe41c8a14e-kube-api-access-hdhk9\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.636818 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.636956 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.738424 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.738917 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.739043 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhk9\" (UniqueName: \"kubernetes.io/projected/a093384f-2994-4207-8447-1cfe41c8a14e-kube-api-access-hdhk9\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.739150 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.744757 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.745045 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.752106 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a093384f-2994-4207-8447-1cfe41c8a14e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.761264 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhk9\" (UniqueName: \"kubernetes.io/projected/a093384f-2994-4207-8447-1cfe41c8a14e-kube-api-access-hdhk9\") pod \"kube-state-metrics-0\" (UID: \"a093384f-2994-4207-8447-1cfe41c8a14e\") " pod="openstack/kube-state-metrics-0" Dec 02 13:24:47 crc kubenswrapper[4725]: I1202 13:24:47.927400 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.444670 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 13:24:48 crc kubenswrapper[4725]: W1202 13:24:48.447167 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda093384f_2994_4207_8447_1cfe41c8a14e.slice/crio-ad820af67473ce668c47b7fa1f25c2597a204edbb62bf2b0d6261fb2bbdeb7d7 WatchSource:0}: Error finding container ad820af67473ce668c47b7fa1f25c2597a204edbb62bf2b0d6261fb2bbdeb7d7: Status 404 returned error can't find the container with id ad820af67473ce668c47b7fa1f25c2597a204edbb62bf2b0d6261fb2bbdeb7d7 Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.536680 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a093384f-2994-4207-8447-1cfe41c8a14e","Type":"ContainerStarted","Data":"ad820af67473ce668c47b7fa1f25c2597a204edbb62bf2b0d6261fb2bbdeb7d7"} Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.677452 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.677803 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-central-agent" containerID="cri-o://0ac6b72baf26b034e2331ed06d4e6f8dc65fd872083df74a82bc77772f07f482" gracePeriod=30 Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.677868 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-notification-agent" containerID="cri-o://5381c380d7ff1a0de5c1da58b660a78160743027d1ff9ce611c0682fbe914dd3" gracePeriod=30 Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.677881 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="sg-core" containerID="cri-o://bd5b83b7bb70d4c58e86a17406f4109f74050af8c069b943d87ddec9083c52da" gracePeriod=30 Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.677885 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="proxy-httpd" containerID="cri-o://62e4b0bbe0a93f4db166ad5185700e98ab470446aa586e83b5af44ca3f792b85" gracePeriod=30 Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.858080 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.901435 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.946608 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 13:24:48 crc kubenswrapper[4725]: I1202 13:24:48.946645 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.052892 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.077613 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.145984 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-xtptm"] Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.146204 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" containerName="dnsmasq-dns" containerID="cri-o://5bc681abf46687702379f1e084a10192f803233e7ce723b8f9138dde893032de" gracePeriod=10 Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.318654 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bbe1936-05fb-4f72-affd-7573be426c6e" path="/var/lib/kubelet/pods/1bbe1936-05fb-4f72-affd-7573be426c6e/volumes" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.559061 4725 generic.go:334] "Generic (PLEG): container finished" podID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerID="62e4b0bbe0a93f4db166ad5185700e98ab470446aa586e83b5af44ca3f792b85" exitCode=0 Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.559109 4725 generic.go:334] "Generic (PLEG): container finished" podID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerID="bd5b83b7bb70d4c58e86a17406f4109f74050af8c069b943d87ddec9083c52da" exitCode=2 Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.559120 4725 generic.go:334] "Generic (PLEG): container finished" podID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerID="0ac6b72baf26b034e2331ed06d4e6f8dc65fd872083df74a82bc77772f07f482" exitCode=0 Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.559168 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerDied","Data":"62e4b0bbe0a93f4db166ad5185700e98ab470446aa586e83b5af44ca3f792b85"} Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.559204 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerDied","Data":"bd5b83b7bb70d4c58e86a17406f4109f74050af8c069b943d87ddec9083c52da"} Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.559219 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerDied","Data":"0ac6b72baf26b034e2331ed06d4e6f8dc65fd872083df74a82bc77772f07f482"} Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.565247 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a093384f-2994-4207-8447-1cfe41c8a14e","Type":"ContainerStarted","Data":"ed1d673c9eb842e2af69b8bb1de9625dc3da39a3fd1d470fb8f96f11aa1590dc"} Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.565510 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.577363 4725 generic.go:334] "Generic (PLEG): container finished" podID="0e621595-2643-4b38-913c-23bc8f80fb62" containerID="5bc681abf46687702379f1e084a10192f803233e7ce723b8f9138dde893032de" exitCode=0 Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.577435 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" event={"ID":"0e621595-2643-4b38-913c-23bc8f80fb62","Type":"ContainerDied","Data":"5bc681abf46687702379f1e084a10192f803233e7ce723b8f9138dde893032de"} Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.579456 4725 generic.go:334] "Generic (PLEG): container finished" podID="6f6eb9bb-6946-4475-9421-8ef60d3cb50b" containerID="2f7f59366c6df9411bd2995749cd6abbc1d98a0170ea697ab1955c05a1b0f28e" exitCode=0 Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.579592 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cltvz" event={"ID":"6f6eb9bb-6946-4475-9421-8ef60d3cb50b","Type":"ContainerDied","Data":"2f7f59366c6df9411bd2995749cd6abbc1d98a0170ea697ab1955c05a1b0f28e"} Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.593284 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.216421142 podStartE2EDuration="2.593265612s" podCreationTimestamp="2025-12-02 13:24:47 +0000 UTC" firstStartedPulling="2025-12-02 13:24:48.450950917 +0000 UTC m=+1219.407592622" lastFinishedPulling="2025-12-02 13:24:48.827795397 +0000 UTC m=+1219.784437092" observedRunningTime="2025-12-02 13:24:49.589938138 +0000 UTC m=+1220.546579833" watchObservedRunningTime="2025-12-02 13:24:49.593265612 +0000 UTC m=+1220.549907307" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.629731 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.887560 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.892135 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:24:49 crc kubenswrapper[4725]: I1202 13:24:49.894158 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.031750 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.032124 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.096166 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j6jv\" (UniqueName: \"kubernetes.io/projected/0e621595-2643-4b38-913c-23bc8f80fb62-kube-api-access-2j6jv\") pod \"0e621595-2643-4b38-913c-23bc8f80fb62\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.096388 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-sb\") pod \"0e621595-2643-4b38-913c-23bc8f80fb62\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.096433 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-config\") pod \"0e621595-2643-4b38-913c-23bc8f80fb62\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.096535 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-nb\") pod \"0e621595-2643-4b38-913c-23bc8f80fb62\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.096560 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-swift-storage-0\") pod \"0e621595-2643-4b38-913c-23bc8f80fb62\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.096581 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-svc\") pod \"0e621595-2643-4b38-913c-23bc8f80fb62\" (UID: \"0e621595-2643-4b38-913c-23bc8f80fb62\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.105703 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e621595-2643-4b38-913c-23bc8f80fb62-kube-api-access-2j6jv" (OuterVolumeSpecName: "kube-api-access-2j6jv") pod "0e621595-2643-4b38-913c-23bc8f80fb62" (UID: "0e621595-2643-4b38-913c-23bc8f80fb62"). InnerVolumeSpecName "kube-api-access-2j6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.166728 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0e621595-2643-4b38-913c-23bc8f80fb62" (UID: "0e621595-2643-4b38-913c-23bc8f80fb62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.171732 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0e621595-2643-4b38-913c-23bc8f80fb62" (UID: "0e621595-2643-4b38-913c-23bc8f80fb62"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.186898 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0e621595-2643-4b38-913c-23bc8f80fb62" (UID: "0e621595-2643-4b38-913c-23bc8f80fb62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.188263 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0e621595-2643-4b38-913c-23bc8f80fb62" (UID: "0e621595-2643-4b38-913c-23bc8f80fb62"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.203587 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.203645 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.203656 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.203667 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.203681 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j6jv\" (UniqueName: \"kubernetes.io/projected/0e621595-2643-4b38-913c-23bc8f80fb62-kube-api-access-2j6jv\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.208563 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-config" (OuterVolumeSpecName: "config") pod "0e621595-2643-4b38-913c-23bc8f80fb62" (UID: "0e621595-2643-4b38-913c-23bc8f80fb62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.306967 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e621595-2643-4b38-913c-23bc8f80fb62-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.595428 4725 generic.go:334] "Generic (PLEG): container finished" podID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerID="5381c380d7ff1a0de5c1da58b660a78160743027d1ff9ce611c0682fbe914dd3" exitCode=0 Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.595522 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerDied","Data":"5381c380d7ff1a0de5c1da58b660a78160743027d1ff9ce611c0682fbe914dd3"} Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.598392 4725 generic.go:334] "Generic (PLEG): container finished" podID="a606a54a-2bb1-4390-8682-dd7e78b4e955" containerID="db5cff9dcd14e1d46ebb9ccec66be80fb821bdc98fcb6c2d1debf9cff2c21a17" exitCode=0 Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.598489 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-sffb7" event={"ID":"a606a54a-2bb1-4390-8682-dd7e78b4e955","Type":"ContainerDied","Data":"db5cff9dcd14e1d46ebb9ccec66be80fb821bdc98fcb6c2d1debf9cff2c21a17"} Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.605502 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" event={"ID":"0e621595-2643-4b38-913c-23bc8f80fb62","Type":"ContainerDied","Data":"b84c2df8360ffe39f8ecd6030be8ad018cb68e904bafedb2e739d5c318a269ac"} Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.605591 4725 scope.go:117] "RemoveContainer" containerID="5bc681abf46687702379f1e084a10192f803233e7ce723b8f9138dde893032de" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.605822 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-xtptm" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.666536 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-xtptm"] Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.683315 4725 scope.go:117] "RemoveContainer" containerID="c1f1baa0e09625e9962a5936162f32f51cd35eb41b7296e241cc5250d3f005b0" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.685021 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-xtptm"] Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.809137 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.920062 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-combined-ca-bundle\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.920297 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7c28\" (UniqueName: \"kubernetes.io/projected/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-kube-api-access-m7c28\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.920344 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-run-httpd\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.920652 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-log-httpd\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.920775 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-scripts\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.920849 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-sg-core-conf-yaml\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.921103 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-config-data\") pod \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\" (UID: \"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5\") " Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.921341 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.922158 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.924016 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.928879 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-scripts" (OuterVolumeSpecName: "scripts") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.930204 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-kube-api-access-m7c28" (OuterVolumeSpecName: "kube-api-access-m7c28") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "kube-api-access-m7c28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:50 crc kubenswrapper[4725]: I1202 13:24:50.955911 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.019955 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.025635 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.025684 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.025703 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.025718 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7c28\" (UniqueName: \"kubernetes.io/projected/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-kube-api-access-m7c28\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.025732 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.039346 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.046212 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-config-data" (OuterVolumeSpecName: "config-data") pod "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" (UID: "6e01cbd5-b76f-4e54-bd78-8036f9c69ef5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.131030 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgftn\" (UniqueName: \"kubernetes.io/projected/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-kube-api-access-dgftn\") pod \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.131311 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-combined-ca-bundle\") pod \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.131394 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-scripts\") pod \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.131485 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-config-data\") pod \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\" (UID: \"6f6eb9bb-6946-4475-9421-8ef60d3cb50b\") " Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.132168 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.135230 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-scripts" (OuterVolumeSpecName: "scripts") pod "6f6eb9bb-6946-4475-9421-8ef60d3cb50b" (UID: "6f6eb9bb-6946-4475-9421-8ef60d3cb50b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.135282 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-kube-api-access-dgftn" (OuterVolumeSpecName: "kube-api-access-dgftn") pod "6f6eb9bb-6946-4475-9421-8ef60d3cb50b" (UID: "6f6eb9bb-6946-4475-9421-8ef60d3cb50b"). InnerVolumeSpecName "kube-api-access-dgftn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.165331 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-config-data" (OuterVolumeSpecName: "config-data") pod "6f6eb9bb-6946-4475-9421-8ef60d3cb50b" (UID: "6f6eb9bb-6946-4475-9421-8ef60d3cb50b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.202797 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f6eb9bb-6946-4475-9421-8ef60d3cb50b" (UID: "6f6eb9bb-6946-4475-9421-8ef60d3cb50b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.234403 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.234545 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.234561 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgftn\" (UniqueName: \"kubernetes.io/projected/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-kube-api-access-dgftn\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.234576 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6eb9bb-6946-4475-9421-8ef60d3cb50b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.283698 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" path="/var/lib/kubelet/pods/0e621595-2643-4b38-913c-23bc8f80fb62/volumes" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.649486 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-cltvz" event={"ID":"6f6eb9bb-6946-4475-9421-8ef60d3cb50b","Type":"ContainerDied","Data":"abc3e201017f0cb80df1a48e135a94cee9546d9df744e405cb928ebf0fc811cd"} Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.649888 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abc3e201017f0cb80df1a48e135a94cee9546d9df744e405cb928ebf0fc811cd" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.649648 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-cltvz" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.660641 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.660693 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6e01cbd5-b76f-4e54-bd78-8036f9c69ef5","Type":"ContainerDied","Data":"982cdb17a7d9bae4fc46057b8c0d934cd8a31656bbd37e3a9599bdb6cb1f0c09"} Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.660734 4725 scope.go:117] "RemoveContainer" containerID="62e4b0bbe0a93f4db166ad5185700e98ab470446aa586e83b5af44ca3f792b85" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.688128 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.692634 4725 scope.go:117] "RemoveContainer" containerID="bd5b83b7bb70d4c58e86a17406f4109f74050af8c069b943d87ddec9083c52da" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.697368 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712229 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712768 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" containerName="init" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712792 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" containerName="init" Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712833 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f6eb9bb-6946-4475-9421-8ef60d3cb50b" containerName="nova-manage" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712842 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f6eb9bb-6946-4475-9421-8ef60d3cb50b" containerName="nova-manage" Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712854 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="proxy-httpd" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712861 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="proxy-httpd" Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712876 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-notification-agent" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712883 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-notification-agent" Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712893 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="sg-core" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712900 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="sg-core" Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712909 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" containerName="dnsmasq-dns" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712917 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" containerName="dnsmasq-dns" Dec 02 13:24:51 crc kubenswrapper[4725]: E1202 13:24:51.712937 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-central-agent" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.712947 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-central-agent" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.713152 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f6eb9bb-6946-4475-9421-8ef60d3cb50b" containerName="nova-manage" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.713172 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="proxy-httpd" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.713184 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-central-agent" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.713200 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e621595-2643-4b38-913c-23bc8f80fb62" containerName="dnsmasq-dns" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.713213 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="sg-core" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.713235 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" containerName="ceilometer-notification-agent" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.715414 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.725162 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.725253 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.725407 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.728227 4725 scope.go:117] "RemoveContainer" containerID="5381c380d7ff1a0de5c1da58b660a78160743027d1ff9ce611c0682fbe914dd3" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.726990 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760337 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-run-httpd\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760508 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760539 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8vjz\" (UniqueName: \"kubernetes.io/projected/53e3272a-6a56-4bba-8a07-48997cfc1dc8-kube-api-access-f8vjz\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760610 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-config-data\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760641 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-log-httpd\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760698 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.760765 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-scripts\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.761094 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.777675 4725 scope.go:117] "RemoveContainer" containerID="0ac6b72baf26b034e2331ed06d4e6f8dc65fd872083df74a82bc77772f07f482" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.845173 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.845425 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-log" containerID="cri-o://42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330" gracePeriod=30 Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.846005 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-api" containerID="cri-o://540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23" gracePeriod=30 Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.862827 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-config-data\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.862865 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-log-httpd\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.862900 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.862923 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-scripts\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.862979 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.863048 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-run-httpd\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.863091 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.863119 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8vjz\" (UniqueName: \"kubernetes.io/projected/53e3272a-6a56-4bba-8a07-48997cfc1dc8-kube-api-access-f8vjz\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.864557 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-run-httpd\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.867614 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-log-httpd\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.872308 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.873664 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.875665 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-config-data\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.882575 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.886120 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-scripts\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.886170 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.886374 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" containerName="nova-scheduler-scheduler" containerID="cri-o://2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c" gracePeriod=30 Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.900948 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.901401 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-log" containerID="cri-o://6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0" gracePeriod=30 Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.902203 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-metadata" containerID="cri-o://4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a" gracePeriod=30 Dec 02 13:24:51 crc kubenswrapper[4725]: I1202 13:24:51.905187 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8vjz\" (UniqueName: \"kubernetes.io/projected/53e3272a-6a56-4bba-8a07-48997cfc1dc8-kube-api-access-f8vjz\") pod \"ceilometer-0\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " pod="openstack/ceilometer-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.044105 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.049361 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.069037 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-config-data\") pod \"a606a54a-2bb1-4390-8682-dd7e78b4e955\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.069184 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-scripts\") pod \"a606a54a-2bb1-4390-8682-dd7e78b4e955\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.069211 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-combined-ca-bundle\") pod \"a606a54a-2bb1-4390-8682-dd7e78b4e955\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.069287 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhsqr\" (UniqueName: \"kubernetes.io/projected/a606a54a-2bb1-4390-8682-dd7e78b4e955-kube-api-access-hhsqr\") pod \"a606a54a-2bb1-4390-8682-dd7e78b4e955\" (UID: \"a606a54a-2bb1-4390-8682-dd7e78b4e955\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.080339 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-scripts" (OuterVolumeSpecName: "scripts") pod "a606a54a-2bb1-4390-8682-dd7e78b4e955" (UID: "a606a54a-2bb1-4390-8682-dd7e78b4e955"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.080488 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a606a54a-2bb1-4390-8682-dd7e78b4e955-kube-api-access-hhsqr" (OuterVolumeSpecName: "kube-api-access-hhsqr") pod "a606a54a-2bb1-4390-8682-dd7e78b4e955" (UID: "a606a54a-2bb1-4390-8682-dd7e78b4e955"). InnerVolumeSpecName "kube-api-access-hhsqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.102482 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a606a54a-2bb1-4390-8682-dd7e78b4e955" (UID: "a606a54a-2bb1-4390-8682-dd7e78b4e955"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.107568 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-config-data" (OuterVolumeSpecName: "config-data") pod "a606a54a-2bb1-4390-8682-dd7e78b4e955" (UID: "a606a54a-2bb1-4390-8682-dd7e78b4e955"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.171316 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.171863 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.171882 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhsqr\" (UniqueName: \"kubernetes.io/projected/a606a54a-2bb1-4390-8682-dd7e78b4e955-kube-api-access-hhsqr\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.171893 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a606a54a-2bb1-4390-8682-dd7e78b4e955-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.540944 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.610162 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.683410 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-nova-metadata-tls-certs\") pod \"ec1f227c-8fae-46fc-9a85-921c838c9af3\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.683571 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-config-data\") pod \"ec1f227c-8fae-46fc-9a85-921c838c9af3\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.683663 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1f227c-8fae-46fc-9a85-921c838c9af3-logs\") pod \"ec1f227c-8fae-46fc-9a85-921c838c9af3\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.683890 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn5zz\" (UniqueName: \"kubernetes.io/projected/ec1f227c-8fae-46fc-9a85-921c838c9af3-kube-api-access-fn5zz\") pod \"ec1f227c-8fae-46fc-9a85-921c838c9af3\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.684260 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-combined-ca-bundle\") pod \"ec1f227c-8fae-46fc-9a85-921c838c9af3\" (UID: \"ec1f227c-8fae-46fc-9a85-921c838c9af3\") " Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.687113 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec1f227c-8fae-46fc-9a85-921c838c9af3-logs" (OuterVolumeSpecName: "logs") pod "ec1f227c-8fae-46fc-9a85-921c838c9af3" (UID: "ec1f227c-8fae-46fc-9a85-921c838c9af3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.688726 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-sffb7" event={"ID":"a606a54a-2bb1-4390-8682-dd7e78b4e955","Type":"ContainerDied","Data":"2bafcec17c298286035edb8a56bef66e6bd6149bf18e637a65dc563ff297ac67"} Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.688800 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bafcec17c298286035edb8a56bef66e6bd6149bf18e637a65dc563ff297ac67" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.688905 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-sffb7" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.717833 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec1f227c-8fae-46fc-9a85-921c838c9af3-kube-api-access-fn5zz" (OuterVolumeSpecName: "kube-api-access-fn5zz") pod "ec1f227c-8fae-46fc-9a85-921c838c9af3" (UID: "ec1f227c-8fae-46fc-9a85-921c838c9af3"). InnerVolumeSpecName "kube-api-access-fn5zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723045 4725 generic.go:334] "Generic (PLEG): container finished" podID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerID="4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a" exitCode=0 Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723090 4725 generic.go:334] "Generic (PLEG): container finished" podID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerID="6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0" exitCode=143 Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723268 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723715 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1f227c-8fae-46fc-9a85-921c838c9af3","Type":"ContainerDied","Data":"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a"} Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723766 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1f227c-8fae-46fc-9a85-921c838c9af3","Type":"ContainerDied","Data":"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0"} Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723781 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1f227c-8fae-46fc-9a85-921c838c9af3","Type":"ContainerDied","Data":"4ed8c257032687be7f0f5557a08a318adebad6841710ec3cde9d940b4c1708e2"} Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.723798 4725 scope.go:117] "RemoveContainer" containerID="4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.736296 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 13:24:52 crc kubenswrapper[4725]: E1202 13:24:52.736799 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-log" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.736870 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-log" Dec 02 13:24:52 crc kubenswrapper[4725]: E1202 13:24:52.736946 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a606a54a-2bb1-4390-8682-dd7e78b4e955" containerName="nova-cell1-conductor-db-sync" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.737031 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a606a54a-2bb1-4390-8682-dd7e78b4e955" containerName="nova-cell1-conductor-db-sync" Dec 02 13:24:52 crc kubenswrapper[4725]: E1202 13:24:52.737108 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-metadata" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.737166 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-metadata" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.737410 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="a606a54a-2bb1-4390-8682-dd7e78b4e955" containerName="nova-cell1-conductor-db-sync" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.737491 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-log" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.737568 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" containerName="nova-metadata-metadata" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.738280 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerStarted","Data":"0957e7b3ebe40006592225a707ab25e0b9fd72246e2efc4f21771d0a01063d32"} Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.738446 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.741507 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.757756 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.759193 4725 generic.go:334] "Generic (PLEG): container finished" podID="2c05245b-3599-4732-a65e-cbfda339ca53" containerID="42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330" exitCode=143 Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.759285 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c05245b-3599-4732-a65e-cbfda339ca53","Type":"ContainerDied","Data":"42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330"} Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.768149 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-config-data" (OuterVolumeSpecName: "config-data") pod "ec1f227c-8fae-46fc-9a85-921c838c9af3" (UID: "ec1f227c-8fae-46fc-9a85-921c838c9af3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.788231 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55mdd\" (UniqueName: \"kubernetes.io/projected/71998d01-30ab-421a-bddc-5e49c6ac8b29-kube-api-access-55mdd\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.788293 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71998d01-30ab-421a-bddc-5e49c6ac8b29-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.788355 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71998d01-30ab-421a-bddc-5e49c6ac8b29-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.788432 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.788443 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1f227c-8fae-46fc-9a85-921c838c9af3-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.788452 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn5zz\" (UniqueName: \"kubernetes.io/projected/ec1f227c-8fae-46fc-9a85-921c838c9af3-kube-api-access-fn5zz\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.804513 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ec1f227c-8fae-46fc-9a85-921c838c9af3" (UID: "ec1f227c-8fae-46fc-9a85-921c838c9af3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.805311 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec1f227c-8fae-46fc-9a85-921c838c9af3" (UID: "ec1f227c-8fae-46fc-9a85-921c838c9af3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.807613 4725 scope.go:117] "RemoveContainer" containerID="6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.835838 4725 scope.go:117] "RemoveContainer" containerID="4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a" Dec 02 13:24:52 crc kubenswrapper[4725]: E1202 13:24:52.836336 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a\": container with ID starting with 4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a not found: ID does not exist" containerID="4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.836381 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a"} err="failed to get container status \"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a\": rpc error: code = NotFound desc = could not find container \"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a\": container with ID starting with 4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a not found: ID does not exist" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.836409 4725 scope.go:117] "RemoveContainer" containerID="6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0" Dec 02 13:24:52 crc kubenswrapper[4725]: E1202 13:24:52.845953 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0\": container with ID starting with 6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0 not found: ID does not exist" containerID="6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.846037 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0"} err="failed to get container status \"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0\": rpc error: code = NotFound desc = could not find container \"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0\": container with ID starting with 6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0 not found: ID does not exist" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.846075 4725 scope.go:117] "RemoveContainer" containerID="4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.847221 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a"} err="failed to get container status \"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a\": rpc error: code = NotFound desc = could not find container \"4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a\": container with ID starting with 4d13563c5bda10d86e6371ac1f4292d12c8958fd9214c6f9ae67dedd6510075a not found: ID does not exist" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.847252 4725 scope.go:117] "RemoveContainer" containerID="6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.847588 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0"} err="failed to get container status \"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0\": rpc error: code = NotFound desc = could not find container \"6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0\": container with ID starting with 6e924d302dae99ccb436e22f0ee1d232abce4ae1b22dcadf09bbe1fea88a32b0 not found: ID does not exist" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.889783 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55mdd\" (UniqueName: \"kubernetes.io/projected/71998d01-30ab-421a-bddc-5e49c6ac8b29-kube-api-access-55mdd\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.890136 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71998d01-30ab-421a-bddc-5e49c6ac8b29-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.890308 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71998d01-30ab-421a-bddc-5e49c6ac8b29-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.890606 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.890722 4725 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec1f227c-8fae-46fc-9a85-921c838c9af3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.894059 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71998d01-30ab-421a-bddc-5e49c6ac8b29-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.894737 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71998d01-30ab-421a-bddc-5e49c6ac8b29-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:52 crc kubenswrapper[4725]: I1202 13:24:52.907733 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55mdd\" (UniqueName: \"kubernetes.io/projected/71998d01-30ab-421a-bddc-5e49c6ac8b29-kube-api-access-55mdd\") pod \"nova-cell1-conductor-0\" (UID: \"71998d01-30ab-421a-bddc-5e49c6ac8b29\") " pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.067737 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.078905 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.091115 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.092654 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.096376 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.096389 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.106324 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.107914 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.197138 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-config-data\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.197722 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.197863 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjhl8\" (UniqueName: \"kubernetes.io/projected/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-kube-api-access-pjhl8\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.197954 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.197989 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-logs\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.300954 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-config-data\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.301064 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.301105 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjhl8\" (UniqueName: \"kubernetes.io/projected/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-kube-api-access-pjhl8\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.301197 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.301258 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-logs\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.301981 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-logs\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.310862 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e01cbd5-b76f-4e54-bd78-8036f9c69ef5" path="/var/lib/kubelet/pods/6e01cbd5-b76f-4e54-bd78-8036f9c69ef5/volumes" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.312377 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec1f227c-8fae-46fc-9a85-921c838c9af3" path="/var/lib/kubelet/pods/ec1f227c-8fae-46fc-9a85-921c838c9af3/volumes" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.313776 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-config-data\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.314862 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.324172 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjhl8\" (UniqueName: \"kubernetes.io/projected/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-kube-api-access-pjhl8\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.333318 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.507327 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.629637 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 13:24:53 crc kubenswrapper[4725]: I1202 13:24:53.900929 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"71998d01-30ab-421a-bddc-5e49c6ac8b29","Type":"ContainerStarted","Data":"92a362e82cbb3f6c929386df9f5120789ada2db29d147f27731ef335572eba60"} Dec 02 13:24:53 crc kubenswrapper[4725]: E1202 13:24:53.901278 4725 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 13:24:53 crc kubenswrapper[4725]: E1202 13:24:53.909614 4725 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 13:24:53 crc kubenswrapper[4725]: E1202 13:24:53.916632 4725 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 13:24:53 crc kubenswrapper[4725]: E1202 13:24:53.916716 4725 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" containerName="nova-scheduler-scheduler" Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.213725 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.919721 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerStarted","Data":"834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462"} Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.926813 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"15e51fdb-0e7a-46e1-959f-b384ebfe25bd","Type":"ContainerStarted","Data":"632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637"} Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.926885 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"15e51fdb-0e7a-46e1-959f-b384ebfe25bd","Type":"ContainerStarted","Data":"771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5"} Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.927089 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"15e51fdb-0e7a-46e1-959f-b384ebfe25bd","Type":"ContainerStarted","Data":"3a491a180a2751841c4ac8caf628414c627ec52118ae6542511ab7d1622bdaa7"} Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.936300 4725 generic.go:334] "Generic (PLEG): container finished" podID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" containerID="2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c" exitCode=0 Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.936432 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd7fc68c-2871-4b3f-96bd-50d9274853e5","Type":"ContainerDied","Data":"2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c"} Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.939206 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"71998d01-30ab-421a-bddc-5e49c6ac8b29","Type":"ContainerStarted","Data":"d4b86975eb5fc47e2cfb4f7d3fc907586cd30dd286eabd0cce80aa96a30c2b33"} Dec 02 13:24:54 crc kubenswrapper[4725]: I1202 13:24:54.940184 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.022648 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.022618015 podStartE2EDuration="2.022618015s" podCreationTimestamp="2025-12-02 13:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:55.008113202 +0000 UTC m=+1225.964754897" watchObservedRunningTime="2025-12-02 13:24:55.022618015 +0000 UTC m=+1225.979259720" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.050426 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.05039428 podStartE2EDuration="3.05039428s" podCreationTimestamp="2025-12-02 13:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:55.050285907 +0000 UTC m=+1226.006927612" watchObservedRunningTime="2025-12-02 13:24:55.05039428 +0000 UTC m=+1226.007035975" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.313145 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.460070 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-config-data\") pod \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.460125 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle\") pod \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.460185 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjdkg\" (UniqueName: \"kubernetes.io/projected/cd7fc68c-2871-4b3f-96bd-50d9274853e5-kube-api-access-pjdkg\") pod \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.474348 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd7fc68c-2871-4b3f-96bd-50d9274853e5-kube-api-access-pjdkg" (OuterVolumeSpecName: "kube-api-access-pjdkg") pod "cd7fc68c-2871-4b3f-96bd-50d9274853e5" (UID: "cd7fc68c-2871-4b3f-96bd-50d9274853e5"). InnerVolumeSpecName "kube-api-access-pjdkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:55 crc kubenswrapper[4725]: E1202 13:24:55.496556 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle podName:cd7fc68c-2871-4b3f-96bd-50d9274853e5 nodeName:}" failed. No retries permitted until 2025-12-02 13:24:55.996512083 +0000 UTC m=+1226.953153768 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle") pod "cd7fc68c-2871-4b3f-96bd-50d9274853e5" (UID: "cd7fc68c-2871-4b3f-96bd-50d9274853e5") : error deleting /var/lib/kubelet/pods/cd7fc68c-2871-4b3f-96bd-50d9274853e5/volume-subpaths: remove /var/lib/kubelet/pods/cd7fc68c-2871-4b3f-96bd-50d9274853e5/volume-subpaths: no such file or directory Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.500450 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-config-data" (OuterVolumeSpecName: "config-data") pod "cd7fc68c-2871-4b3f-96bd-50d9274853e5" (UID: "cd7fc68c-2871-4b3f-96bd-50d9274853e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.563164 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.563221 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjdkg\" (UniqueName: \"kubernetes.io/projected/cd7fc68c-2871-4b3f-96bd-50d9274853e5-kube-api-access-pjdkg\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.951156 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cd7fc68c-2871-4b3f-96bd-50d9274853e5","Type":"ContainerDied","Data":"22730c89e0ec25818b29cf9db3ca6742f62999beb6c4e2f97dda0585d7103b0f"} Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.951199 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.951232 4725 scope.go:117] "RemoveContainer" containerID="2f8bf9b0ef2e2364ac1c26d15205131f505da36213b925c2846b584a930c441c" Dec 02 13:24:55 crc kubenswrapper[4725]: I1202 13:24:55.953593 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerStarted","Data":"97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f"} Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.075943 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle\") pod \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\" (UID: \"cd7fc68c-2871-4b3f-96bd-50d9274853e5\") " Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.080725 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd7fc68c-2871-4b3f-96bd-50d9274853e5" (UID: "cd7fc68c-2871-4b3f-96bd-50d9274853e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.180027 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fc68c-2871-4b3f-96bd-50d9274853e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.292340 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.311513 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.326373 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:56 crc kubenswrapper[4725]: E1202 13:24:56.327071 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" containerName="nova-scheduler-scheduler" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.327099 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" containerName="nova-scheduler-scheduler" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.327297 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" containerName="nova-scheduler-scheduler" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.328439 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.330561 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.341722 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.486919 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-config-data\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.486980 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.487026 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7fxh\" (UniqueName: \"kubernetes.io/projected/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-kube-api-access-v7fxh\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.589093 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-config-data\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.589174 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.589237 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7fxh\" (UniqueName: \"kubernetes.io/projected/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-kube-api-access-v7fxh\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.601061 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.603022 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-config-data\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.616922 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7fxh\" (UniqueName: \"kubernetes.io/projected/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-kube-api-access-v7fxh\") pod \"nova-scheduler-0\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.654234 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.783925 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.897680 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-config-data\") pod \"2c05245b-3599-4732-a65e-cbfda339ca53\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.897771 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c05245b-3599-4732-a65e-cbfda339ca53-logs\") pod \"2c05245b-3599-4732-a65e-cbfda339ca53\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.897827 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9xbz\" (UniqueName: \"kubernetes.io/projected/2c05245b-3599-4732-a65e-cbfda339ca53-kube-api-access-c9xbz\") pod \"2c05245b-3599-4732-a65e-cbfda339ca53\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.898085 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-combined-ca-bundle\") pod \"2c05245b-3599-4732-a65e-cbfda339ca53\" (UID: \"2c05245b-3599-4732-a65e-cbfda339ca53\") " Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.899674 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c05245b-3599-4732-a65e-cbfda339ca53-logs" (OuterVolumeSpecName: "logs") pod "2c05245b-3599-4732-a65e-cbfda339ca53" (UID: "2c05245b-3599-4732-a65e-cbfda339ca53"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.915184 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c05245b-3599-4732-a65e-cbfda339ca53-kube-api-access-c9xbz" (OuterVolumeSpecName: "kube-api-access-c9xbz") pod "2c05245b-3599-4732-a65e-cbfda339ca53" (UID: "2c05245b-3599-4732-a65e-cbfda339ca53"). InnerVolumeSpecName "kube-api-access-c9xbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.938384 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-config-data" (OuterVolumeSpecName: "config-data") pod "2c05245b-3599-4732-a65e-cbfda339ca53" (UID: "2c05245b-3599-4732-a65e-cbfda339ca53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.942803 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c05245b-3599-4732-a65e-cbfda339ca53" (UID: "2c05245b-3599-4732-a65e-cbfda339ca53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.973868 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerStarted","Data":"91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3"} Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.977032 4725 generic.go:334] "Generic (PLEG): container finished" podID="2c05245b-3599-4732-a65e-cbfda339ca53" containerID="540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23" exitCode=0 Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.977116 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.977100 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c05245b-3599-4732-a65e-cbfda339ca53","Type":"ContainerDied","Data":"540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23"} Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.977185 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2c05245b-3599-4732-a65e-cbfda339ca53","Type":"ContainerDied","Data":"0a701eedba18918b4edab06f16d52f59c3161f3c552b514db2a5eef61af3906b"} Dec 02 13:24:56 crc kubenswrapper[4725]: I1202 13:24:56.977214 4725 scope.go:117] "RemoveContainer" containerID="540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.001284 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.001317 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c05245b-3599-4732-a65e-cbfda339ca53-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.001329 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c05245b-3599-4732-a65e-cbfda339ca53-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.001338 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9xbz\" (UniqueName: \"kubernetes.io/projected/2c05245b-3599-4732-a65e-cbfda339ca53-kube-api-access-c9xbz\") on node \"crc\" DevicePath \"\"" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.019328 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.025524 4725 scope.go:117] "RemoveContainer" containerID="42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.048973 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.061470 4725 scope.go:117] "RemoveContainer" containerID="540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23" Dec 02 13:24:57 crc kubenswrapper[4725]: E1202 13:24:57.062082 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23\": container with ID starting with 540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23 not found: ID does not exist" containerID="540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.062164 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23"} err="failed to get container status \"540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23\": rpc error: code = NotFound desc = could not find container \"540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23\": container with ID starting with 540385d6046d5daa95082321a4833ad03675c86b6ceaa6bafde046d80166ee23 not found: ID does not exist" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.062210 4725 scope.go:117] "RemoveContainer" containerID="42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330" Dec 02 13:24:57 crc kubenswrapper[4725]: E1202 13:24:57.062613 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330\": container with ID starting with 42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330 not found: ID does not exist" containerID="42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.062658 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330"} err="failed to get container status \"42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330\": rpc error: code = NotFound desc = could not find container \"42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330\": container with ID starting with 42b01cf8b9899b132ff7ba61b442b6d5f7576ed8923d5dee4e87b8f526624330 not found: ID does not exist" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.062789 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:57 crc kubenswrapper[4725]: E1202 13:24:57.063354 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-log" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.063371 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-log" Dec 02 13:24:57 crc kubenswrapper[4725]: E1202 13:24:57.063381 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-api" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.063387 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-api" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.063594 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-log" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.063629 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" containerName="nova-api-api" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.064871 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.070847 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.076287 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.182854 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.207911 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9bs8\" (UniqueName: \"kubernetes.io/projected/548320bf-6a44-424e-af63-6900934d7ae3-kube-api-access-l9bs8\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.208065 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548320bf-6a44-424e-af63-6900934d7ae3-logs\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.208096 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-config-data\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.208112 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.283526 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c05245b-3599-4732-a65e-cbfda339ca53" path="/var/lib/kubelet/pods/2c05245b-3599-4732-a65e-cbfda339ca53/volumes" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.284600 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd7fc68c-2871-4b3f-96bd-50d9274853e5" path="/var/lib/kubelet/pods/cd7fc68c-2871-4b3f-96bd-50d9274853e5/volumes" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.309685 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548320bf-6a44-424e-af63-6900934d7ae3-logs\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.309735 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-config-data\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.309755 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.309854 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9bs8\" (UniqueName: \"kubernetes.io/projected/548320bf-6a44-424e-af63-6900934d7ae3-kube-api-access-l9bs8\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.310222 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548320bf-6a44-424e-af63-6900934d7ae3-logs\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.314301 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-config-data\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.319276 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.335167 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9bs8\" (UniqueName: \"kubernetes.io/projected/548320bf-6a44-424e-af63-6900934d7ae3-kube-api-access-l9bs8\") pod \"nova-api-0\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.390427 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.901266 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:24:57 crc kubenswrapper[4725]: I1202 13:24:57.944844 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.013680 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerStarted","Data":"ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb"} Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.014130 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.032139 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"548320bf-6a44-424e-af63-6900934d7ae3","Type":"ContainerStarted","Data":"aa6eca6c1721e7af771c735aaa410b6b134f2c7950835b1545bd6566ee164a2b"} Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.033809 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88afcfe6-5dc2-4ffc-a79b-4c46585122bf","Type":"ContainerStarted","Data":"15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f"} Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.033849 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88afcfe6-5dc2-4ffc-a79b-4c46585122bf","Type":"ContainerStarted","Data":"04ebd6aa534134bb21715c3d8c6434a5835e141c4d2544487acfd854d5ebd02d"} Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.050165 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.927723173 podStartE2EDuration="7.050118364s" podCreationTimestamp="2025-12-02 13:24:51 +0000 UTC" firstStartedPulling="2025-12-02 13:24:52.544216496 +0000 UTC m=+1223.500858191" lastFinishedPulling="2025-12-02 13:24:57.666611687 +0000 UTC m=+1228.623253382" observedRunningTime="2025-12-02 13:24:58.046187865 +0000 UTC m=+1229.002829560" watchObservedRunningTime="2025-12-02 13:24:58.050118364 +0000 UTC m=+1229.006760059" Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.084351 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.08432573 podStartE2EDuration="2.08432573s" podCreationTimestamp="2025-12-02 13:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:58.070974155 +0000 UTC m=+1229.027615840" watchObservedRunningTime="2025-12-02 13:24:58.08432573 +0000 UTC m=+1229.040967425" Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.508299 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:24:58 crc kubenswrapper[4725]: I1202 13:24:58.508769 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:24:59 crc kubenswrapper[4725]: I1202 13:24:59.066086 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"548320bf-6a44-424e-af63-6900934d7ae3","Type":"ContainerStarted","Data":"0814506189c624e56d9b7168ab37ee947aae9cd8789af33c2e36de2885ae4b84"} Dec 02 13:24:59 crc kubenswrapper[4725]: I1202 13:24:59.066158 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"548320bf-6a44-424e-af63-6900934d7ae3","Type":"ContainerStarted","Data":"d835b234a2278f03ff4ebd3ff8daf45f7177209a6d72bb07b6e7c3566ecba284"} Dec 02 13:24:59 crc kubenswrapper[4725]: I1202 13:24:59.101509 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.101475603 podStartE2EDuration="2.101475603s" podCreationTimestamp="2025-12-02 13:24:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:24:59.092791145 +0000 UTC m=+1230.049432840" watchObservedRunningTime="2025-12-02 13:24:59.101475603 +0000 UTC m=+1230.058117298" Dec 02 13:25:01 crc kubenswrapper[4725]: I1202 13:25:01.655309 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 13:25:03 crc kubenswrapper[4725]: I1202 13:25:03.144962 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 02 13:25:03 crc kubenswrapper[4725]: I1202 13:25:03.508349 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 13:25:03 crc kubenswrapper[4725]: I1202 13:25:03.508409 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 13:25:04 crc kubenswrapper[4725]: I1202 13:25:04.524767 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:04 crc kubenswrapper[4725]: I1202 13:25:04.524767 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:06 crc kubenswrapper[4725]: I1202 13:25:06.655740 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 13:25:06 crc kubenswrapper[4725]: I1202 13:25:06.680516 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 13:25:07 crc kubenswrapper[4725]: I1202 13:25:07.177321 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 13:25:07 crc kubenswrapper[4725]: I1202 13:25:07.391284 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 13:25:07 crc kubenswrapper[4725]: I1202 13:25:07.391348 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 13:25:08 crc kubenswrapper[4725]: I1202 13:25:08.474744 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:08 crc kubenswrapper[4725]: I1202 13:25:08.474763 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:13 crc kubenswrapper[4725]: I1202 13:25:13.513009 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 13:25:13 crc kubenswrapper[4725]: I1202 13:25:13.513608 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 13:25:13 crc kubenswrapper[4725]: I1202 13:25:13.517553 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 13:25:13 crc kubenswrapper[4725]: I1202 13:25:13.518269 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 13:25:14 crc kubenswrapper[4725]: I1202 13:25:14.959444 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.114337 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-config-data\") pod \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.114651 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxz7c\" (UniqueName: \"kubernetes.io/projected/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-kube-api-access-mxz7c\") pod \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.114682 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-combined-ca-bundle\") pod \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\" (UID: \"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28\") " Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.120249 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-kube-api-access-mxz7c" (OuterVolumeSpecName: "kube-api-access-mxz7c") pod "4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" (UID: "4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28"). InnerVolumeSpecName "kube-api-access-mxz7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.141658 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" (UID: "4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.143318 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-config-data" (OuterVolumeSpecName: "config-data") pod "4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" (UID: "4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218043 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218078 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxz7c\" (UniqueName: \"kubernetes.io/projected/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-kube-api-access-mxz7c\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218114 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218715 4725 generic.go:334] "Generic (PLEG): container finished" podID="4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" containerID="78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757" exitCode=137 Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218812 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218820 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28","Type":"ContainerDied","Data":"78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757"} Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218878 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28","Type":"ContainerDied","Data":"ea5321566e0b5282acba089f46366746ba3aff6eeb43dd0df3634339a8847c20"} Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.218895 4725 scope.go:117] "RemoveContainer" containerID="78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.251059 4725 scope.go:117] "RemoveContainer" containerID="78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757" Dec 02 13:25:15 crc kubenswrapper[4725]: E1202 13:25:15.251566 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757\": container with ID starting with 78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757 not found: ID does not exist" containerID="78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.251598 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757"} err="failed to get container status \"78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757\": rpc error: code = NotFound desc = could not find container \"78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757\": container with ID starting with 78a1a489c0f8bdaf46c50bbb595e0e2e23dd4d9c745d775e7c3caba661044757 not found: ID does not exist" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.252611 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.266478 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.284236 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" path="/var/lib/kubelet/pods/4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28/volumes" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.284883 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:25:15 crc kubenswrapper[4725]: E1202 13:25:15.285253 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.285277 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.285567 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ee4ecec-0147-47a0-bcaa-4f00e0ee5f28" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.286359 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.289135 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.289909 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.290212 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.290353 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.320172 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.320336 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.320427 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrz58\" (UniqueName: \"kubernetes.io/projected/03aa0ff8-1156-466a-9726-48481ee110d8-kube-api-access-rrz58\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.320602 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.320632 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.422745 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.422831 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrz58\" (UniqueName: \"kubernetes.io/projected/03aa0ff8-1156-466a-9726-48481ee110d8-kube-api-access-rrz58\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.422890 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.422910 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.422952 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.426979 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.427249 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.427458 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.427532 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/03aa0ff8-1156-466a-9726-48481ee110d8-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.438057 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrz58\" (UniqueName: \"kubernetes.io/projected/03aa0ff8-1156-466a-9726-48481ee110d8-kube-api-access-rrz58\") pod \"nova-cell1-novncproxy-0\" (UID: \"03aa0ff8-1156-466a-9726-48481ee110d8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:15 crc kubenswrapper[4725]: I1202 13:25:15.607003 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:16 crc kubenswrapper[4725]: I1202 13:25:16.057959 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 13:25:16 crc kubenswrapper[4725]: I1202 13:25:16.234951 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03aa0ff8-1156-466a-9726-48481ee110d8","Type":"ContainerStarted","Data":"b728f47ded17eed9350f5ade11fb21b951b7c958fe953998a45f6059a986acb7"} Dec 02 13:25:17 crc kubenswrapper[4725]: I1202 13:25:17.245656 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03aa0ff8-1156-466a-9726-48481ee110d8","Type":"ContainerStarted","Data":"30a1d0637106a07ca52481788b73bd0eb1b62deacfa4f7cb06ba882ce392791c"} Dec 02 13:25:17 crc kubenswrapper[4725]: I1202 13:25:17.276221 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.27619681 podStartE2EDuration="2.27619681s" podCreationTimestamp="2025-12-02 13:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:17.265902492 +0000 UTC m=+1248.222544197" watchObservedRunningTime="2025-12-02 13:25:17.27619681 +0000 UTC m=+1248.232838525" Dec 02 13:25:17 crc kubenswrapper[4725]: I1202 13:25:17.396307 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 13:25:17 crc kubenswrapper[4725]: I1202 13:25:17.397111 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 13:25:17 crc kubenswrapper[4725]: I1202 13:25:17.397417 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 13:25:17 crc kubenswrapper[4725]: I1202 13:25:17.410566 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.254216 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.258152 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.436722 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-62xsd"] Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.438783 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.451910 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-62xsd"] Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.480662 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.480724 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-config\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.480766 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.480828 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.480875 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.480962 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5lbz\" (UniqueName: \"kubernetes.io/projected/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-kube-api-access-p5lbz\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.582237 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.582300 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-config\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.582336 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.582387 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.582417 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.582497 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5lbz\" (UniqueName: \"kubernetes.io/projected/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-kube-api-access-p5lbz\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.583209 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.583285 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-config\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.584559 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.585886 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.586083 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.608338 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5lbz\" (UniqueName: \"kubernetes.io/projected/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-kube-api-access-p5lbz\") pod \"dnsmasq-dns-89c5cd4d5-62xsd\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:18 crc kubenswrapper[4725]: I1202 13:25:18.763644 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:19 crc kubenswrapper[4725]: I1202 13:25:19.326111 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-62xsd"] Dec 02 13:25:19 crc kubenswrapper[4725]: W1202 13:25:19.341641 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5bec6b9_6b89_4569_a34c_ec75f0027ef6.slice/crio-9dcc42c22d1e51248c995118b51088bc9f20705d8108b8102055f5c1c1e5bad4 WatchSource:0}: Error finding container 9dcc42c22d1e51248c995118b51088bc9f20705d8108b8102055f5c1c1e5bad4: Status 404 returned error can't find the container with id 9dcc42c22d1e51248c995118b51088bc9f20705d8108b8102055f5c1c1e5bad4 Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.277282 4725 generic.go:334] "Generic (PLEG): container finished" podID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerID="2db1493e5faad9e9a103bfee3045f3c3a26f0c1d928a7b6ebe08b895ca9b1dc2" exitCode=0 Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.277368 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" event={"ID":"f5bec6b9-6b89-4569-a34c-ec75f0027ef6","Type":"ContainerDied","Data":"2db1493e5faad9e9a103bfee3045f3c3a26f0c1d928a7b6ebe08b895ca9b1dc2"} Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.277836 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" event={"ID":"f5bec6b9-6b89-4569-a34c-ec75f0027ef6","Type":"ContainerStarted","Data":"9dcc42c22d1e51248c995118b51088bc9f20705d8108b8102055f5c1c1e5bad4"} Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.607315 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.617895 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.618253 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-central-agent" containerID="cri-o://834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462" gracePeriod=30 Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.618333 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-notification-agent" containerID="cri-o://97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f" gracePeriod=30 Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.618348 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="sg-core" containerID="cri-o://91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3" gracePeriod=30 Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.618483 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="proxy-httpd" containerID="cri-o://ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb" gracePeriod=30 Dec 02 13:25:20 crc kubenswrapper[4725]: I1202 13:25:20.630170 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.192:3000/\": read tcp 10.217.0.2:49446->10.217.0.192:3000: read: connection reset by peer" Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.194555 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.289682 4725 generic.go:334] "Generic (PLEG): container finished" podID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerID="ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb" exitCode=0 Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.289744 4725 generic.go:334] "Generic (PLEG): container finished" podID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerID="91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3" exitCode=2 Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.289751 4725 generic.go:334] "Generic (PLEG): container finished" podID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerID="834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462" exitCode=0 Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.289786 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerDied","Data":"ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb"} Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.289811 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerDied","Data":"91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3"} Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.289822 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerDied","Data":"834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462"} Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.291621 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-log" containerID="cri-o://d835b234a2278f03ff4ebd3ff8daf45f7177209a6d72bb07b6e7c3566ecba284" gracePeriod=30 Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.292743 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" event={"ID":"f5bec6b9-6b89-4569-a34c-ec75f0027ef6","Type":"ContainerStarted","Data":"e930f4e708ae92c1a8e99fe114e6486ddfe3d54271084236ff0a69c1c07e4251"} Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.292775 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.293046 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-api" containerID="cri-o://0814506189c624e56d9b7168ab37ee947aae9cd8789af33c2e36de2885ae4b84" gracePeriod=30 Dec 02 13:25:21 crc kubenswrapper[4725]: I1202 13:25:21.321799 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" podStartSLOduration=3.321782777 podStartE2EDuration="3.321782777s" podCreationTimestamp="2025-12-02 13:25:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:21.318823902 +0000 UTC m=+1252.275465597" watchObservedRunningTime="2025-12-02 13:25:21.321782777 +0000 UTC m=+1252.278424462" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.272718 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.306371 4725 generic.go:334] "Generic (PLEG): container finished" podID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerID="97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f" exitCode=0 Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.306439 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerDied","Data":"97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f"} Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.306502 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53e3272a-6a56-4bba-8a07-48997cfc1dc8","Type":"ContainerDied","Data":"0957e7b3ebe40006592225a707ab25e0b9fd72246e2efc4f21771d0a01063d32"} Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.306523 4725 scope.go:117] "RemoveContainer" containerID="ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.306645 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.311241 4725 generic.go:334] "Generic (PLEG): container finished" podID="548320bf-6a44-424e-af63-6900934d7ae3" containerID="d835b234a2278f03ff4ebd3ff8daf45f7177209a6d72bb07b6e7c3566ecba284" exitCode=143 Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.311531 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"548320bf-6a44-424e-af63-6900934d7ae3","Type":"ContainerDied","Data":"d835b234a2278f03ff4ebd3ff8daf45f7177209a6d72bb07b6e7c3566ecba284"} Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.339301 4725 scope.go:117] "RemoveContainer" containerID="91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.363736 4725 scope.go:117] "RemoveContainer" containerID="97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385548 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-sg-core-conf-yaml\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385607 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-combined-ca-bundle\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385637 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-scripts\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385669 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8vjz\" (UniqueName: \"kubernetes.io/projected/53e3272a-6a56-4bba-8a07-48997cfc1dc8-kube-api-access-f8vjz\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385758 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-run-httpd\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385806 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-config-data\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385826 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-log-httpd\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.385961 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-ceilometer-tls-certs\") pod \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\" (UID: \"53e3272a-6a56-4bba-8a07-48997cfc1dc8\") " Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.390705 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.390833 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.396567 4725 scope.go:117] "RemoveContainer" containerID="834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.397853 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e3272a-6a56-4bba-8a07-48997cfc1dc8-kube-api-access-f8vjz" (OuterVolumeSpecName: "kube-api-access-f8vjz") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "kube-api-access-f8vjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.404923 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-scripts" (OuterVolumeSpecName: "scripts") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.437653 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.454981 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.487864 4725 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.487909 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.487926 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.487939 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8vjz\" (UniqueName: \"kubernetes.io/projected/53e3272a-6a56-4bba-8a07-48997cfc1dc8-kube-api-access-f8vjz\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.487950 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.487961 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53e3272a-6a56-4bba-8a07-48997cfc1dc8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.490807 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.530198 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-config-data" (OuterVolumeSpecName: "config-data") pod "53e3272a-6a56-4bba-8a07-48997cfc1dc8" (UID: "53e3272a-6a56-4bba-8a07-48997cfc1dc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.547034 4725 scope.go:117] "RemoveContainer" containerID="ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.547696 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb\": container with ID starting with ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb not found: ID does not exist" containerID="ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.547733 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb"} err="failed to get container status \"ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb\": rpc error: code = NotFound desc = could not find container \"ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb\": container with ID starting with ada08961bfd35b30ce830107835903e3c81b03fb069c15051a1efbada218d7eb not found: ID does not exist" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.547754 4725 scope.go:117] "RemoveContainer" containerID="91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.547956 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3\": container with ID starting with 91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3 not found: ID does not exist" containerID="91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.547979 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3"} err="failed to get container status \"91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3\": rpc error: code = NotFound desc = could not find container \"91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3\": container with ID starting with 91f2f49726944fb259fe6fec2edcc77b22a8821accc029ba136566f6689108a3 not found: ID does not exist" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.547994 4725 scope.go:117] "RemoveContainer" containerID="97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.548188 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f\": container with ID starting with 97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f not found: ID does not exist" containerID="97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.548214 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f"} err="failed to get container status \"97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f\": rpc error: code = NotFound desc = could not find container \"97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f\": container with ID starting with 97914821e9a6b847d95b5a858c7849ce5edaa3f16c6e100000b50b4771f6715f not found: ID does not exist" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.548255 4725 scope.go:117] "RemoveContainer" containerID="834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.548444 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462\": container with ID starting with 834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462 not found: ID does not exist" containerID="834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.548485 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462"} err="failed to get container status \"834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462\": rpc error: code = NotFound desc = could not find container \"834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462\": container with ID starting with 834615d5523f920b3d40d1c8f39a61a292554b9a6b2ecdb0d8d6f228b963b462 not found: ID does not exist" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.589692 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.589726 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e3272a-6a56-4bba-8a07-48997cfc1dc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.646244 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.671894 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.685380 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.685913 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-notification-agent" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.685935 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-notification-agent" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.685953 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="proxy-httpd" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.685960 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="proxy-httpd" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.685988 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-central-agent" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.685995 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-central-agent" Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.686017 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="sg-core" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.686023 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="sg-core" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.686202 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-notification-agent" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.686222 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="proxy-httpd" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.686231 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="ceilometer-central-agent" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.686249 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="sg-core" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.688375 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.693928 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.694002 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.696664 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.699703 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.744512 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:22 crc kubenswrapper[4725]: E1202 13:25:22.745275 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-mgmx4 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="68fddd2f-a572-496f-946c-5c289bb1e22f" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.797833 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgmx4\" (UniqueName: \"kubernetes.io/projected/68fddd2f-a572-496f-946c-5c289bb1e22f-kube-api-access-mgmx4\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.797917 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-scripts\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.797974 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.798034 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-run-httpd\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.798090 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-log-httpd\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.798476 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.798617 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-config-data\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.798739 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.906708 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-config-data\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.906922 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.907673 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgmx4\" (UniqueName: \"kubernetes.io/projected/68fddd2f-a572-496f-946c-5c289bb1e22f-kube-api-access-mgmx4\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.907939 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-scripts\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.908129 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.908325 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-run-httpd\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.908760 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-log-httpd\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.908848 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.909259 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-log-httpd\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.909598 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-run-httpd\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.911384 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-config-data\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.911868 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.912187 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.912420 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-scripts\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.917719 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:22 crc kubenswrapper[4725]: I1202 13:25:22.927634 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgmx4\" (UniqueName: \"kubernetes.io/projected/68fddd2f-a572-496f-946c-5c289bb1e22f-kube-api-access-mgmx4\") pod \"ceilometer-0\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " pod="openstack/ceilometer-0" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.280274 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" path="/var/lib/kubelet/pods/53e3272a-6a56-4bba-8a07-48997cfc1dc8/volumes" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.324926 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.340668 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.518965 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-combined-ca-bundle\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519011 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-run-httpd\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519097 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-config-data\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519196 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-sg-core-conf-yaml\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519219 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgmx4\" (UniqueName: \"kubernetes.io/projected/68fddd2f-a572-496f-946c-5c289bb1e22f-kube-api-access-mgmx4\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519240 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-ceilometer-tls-certs\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519360 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-scripts\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519427 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-log-httpd\") pod \"68fddd2f-a572-496f-946c-5c289bb1e22f\" (UID: \"68fddd2f-a572-496f-946c-5c289bb1e22f\") " Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519557 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.519871 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.525596 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-config-data" (OuterVolumeSpecName: "config-data") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.526499 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-scripts" (OuterVolumeSpecName: "scripts") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.526646 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.527109 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fddd2f-a572-496f-946c-5c289bb1e22f-kube-api-access-mgmx4" (OuterVolumeSpecName: "kube-api-access-mgmx4") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "kube-api-access-mgmx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.530248 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.530302 4725 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.530316 4725 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/68fddd2f-a572-496f-946c-5c289bb1e22f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.530329 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.530341 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgmx4\" (UniqueName: \"kubernetes.io/projected/68fddd2f-a572-496f-946c-5c289bb1e22f-kube-api-access-mgmx4\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.530361 4725 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.537402 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.539281 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "68fddd2f-a572-496f-946c-5c289bb1e22f" (UID: "68fddd2f-a572-496f-946c-5c289bb1e22f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.632547 4725 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:23 crc kubenswrapper[4725]: I1202 13:25:23.632589 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68fddd2f-a572-496f-946c-5c289bb1e22f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.332755 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.412071 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.423682 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.440739 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.444601 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.447810 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.448084 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.450705 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.451893 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.451924 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.451948 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-config-data\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.452008 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-scripts\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.452047 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns2x2\" (UniqueName: \"kubernetes.io/projected/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-kube-api-access-ns2x2\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.452067 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-log-httpd\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.452116 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-run-httpd\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.452141 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.457014 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554274 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-log-httpd\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554346 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns2x2\" (UniqueName: \"kubernetes.io/projected/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-kube-api-access-ns2x2\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554425 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-run-httpd\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554473 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554511 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554535 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554566 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-config-data\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.554634 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-scripts\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.555997 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-log-httpd\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.556559 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-run-httpd\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.563363 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.563867 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-scripts\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.563927 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.566375 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-config-data\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.566681 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.577902 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns2x2\" (UniqueName: \"kubernetes.io/projected/c4e7bb4e-6337-40e2-bea3-ef63e192d2c0-kube-api-access-ns2x2\") pod \"ceilometer-0\" (UID: \"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0\") " pod="openstack/ceilometer-0" Dec 02 13:25:24 crc kubenswrapper[4725]: I1202 13:25:24.768817 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.264361 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.283272 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fddd2f-a572-496f-946c-5c289bb1e22f" path="/var/lib/kubelet/pods/68fddd2f-a572-496f-946c-5c289bb1e22f/volumes" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.345266 4725 generic.go:334] "Generic (PLEG): container finished" podID="548320bf-6a44-424e-af63-6900934d7ae3" containerID="0814506189c624e56d9b7168ab37ee947aae9cd8789af33c2e36de2885ae4b84" exitCode=0 Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.345315 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"548320bf-6a44-424e-af63-6900934d7ae3","Type":"ContainerDied","Data":"0814506189c624e56d9b7168ab37ee947aae9cd8789af33c2e36de2885ae4b84"} Dec 02 13:25:25 crc kubenswrapper[4725]: W1202 13:25:25.374252 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4e7bb4e_6337_40e2_bea3_ef63e192d2c0.slice/crio-dd0a57edc3457b9f0654eeca8b595ebe3f0eafeff585e8a810b61e3c8e25797a WatchSource:0}: Error finding container dd0a57edc3457b9f0654eeca8b595ebe3f0eafeff585e8a810b61e3c8e25797a: Status 404 returned error can't find the container with id dd0a57edc3457b9f0654eeca8b595ebe3f0eafeff585e8a810b61e3c8e25797a Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.608114 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.613931 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.630089 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.683442 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-config-data\") pod \"548320bf-6a44-424e-af63-6900934d7ae3\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.683569 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548320bf-6a44-424e-af63-6900934d7ae3-logs\") pod \"548320bf-6a44-424e-af63-6900934d7ae3\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.683604 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9bs8\" (UniqueName: \"kubernetes.io/projected/548320bf-6a44-424e-af63-6900934d7ae3-kube-api-access-l9bs8\") pod \"548320bf-6a44-424e-af63-6900934d7ae3\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.683658 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-combined-ca-bundle\") pod \"548320bf-6a44-424e-af63-6900934d7ae3\" (UID: \"548320bf-6a44-424e-af63-6900934d7ae3\") " Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.690082 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/548320bf-6a44-424e-af63-6900934d7ae3-logs" (OuterVolumeSpecName: "logs") pod "548320bf-6a44-424e-af63-6900934d7ae3" (UID: "548320bf-6a44-424e-af63-6900934d7ae3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.696729 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548320bf-6a44-424e-af63-6900934d7ae3-kube-api-access-l9bs8" (OuterVolumeSpecName: "kube-api-access-l9bs8") pod "548320bf-6a44-424e-af63-6900934d7ae3" (UID: "548320bf-6a44-424e-af63-6900934d7ae3"). InnerVolumeSpecName "kube-api-access-l9bs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.730972 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "548320bf-6a44-424e-af63-6900934d7ae3" (UID: "548320bf-6a44-424e-af63-6900934d7ae3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.767543 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-config-data" (OuterVolumeSpecName: "config-data") pod "548320bf-6a44-424e-af63-6900934d7ae3" (UID: "548320bf-6a44-424e-af63-6900934d7ae3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.786732 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.786786 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/548320bf-6a44-424e-af63-6900934d7ae3-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.786799 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9bs8\" (UniqueName: \"kubernetes.io/projected/548320bf-6a44-424e-af63-6900934d7ae3-kube-api-access-l9bs8\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:25 crc kubenswrapper[4725]: I1202 13:25:25.786813 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/548320bf-6a44-424e-af63-6900934d7ae3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.357870 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"548320bf-6a44-424e-af63-6900934d7ae3","Type":"ContainerDied","Data":"aa6eca6c1721e7af771c735aaa410b6b134f2c7950835b1545bd6566ee164a2b"} Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.358321 4725 scope.go:117] "RemoveContainer" containerID="0814506189c624e56d9b7168ab37ee947aae9cd8789af33c2e36de2885ae4b84" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.358520 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.361378 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0","Type":"ContainerStarted","Data":"dd0a57edc3457b9f0654eeca8b595ebe3f0eafeff585e8a810b61e3c8e25797a"} Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.386502 4725 scope.go:117] "RemoveContainer" containerID="d835b234a2278f03ff4ebd3ff8daf45f7177209a6d72bb07b6e7c3566ecba284" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.387885 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.420304 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.432349 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.474815 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:26 crc kubenswrapper[4725]: E1202 13:25:26.475280 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-api" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.475292 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-api" Dec 02 13:25:26 crc kubenswrapper[4725]: E1202 13:25:26.475316 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-log" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.475323 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-log" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.479661 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-log" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.479731 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="548320bf-6a44-424e-af63-6900934d7ae3" containerName="nova-api-api" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.495650 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.499394 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.499831 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.514379 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a72f71c-ec9c-4c47-b888-51c35fbae050-logs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.514539 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.514582 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-public-tls-certs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.514615 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-config-data\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.514683 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.514742 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkdm9\" (UniqueName: \"kubernetes.io/projected/6a72f71c-ec9c-4c47-b888-51c35fbae050-kube-api-access-wkdm9\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.535213 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.574127 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.620573 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a72f71c-ec9c-4c47-b888-51c35fbae050-logs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.620674 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.620703 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-public-tls-certs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.620724 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-config-data\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.620758 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.620791 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkdm9\" (UniqueName: \"kubernetes.io/projected/6a72f71c-ec9c-4c47-b888-51c35fbae050-kube-api-access-wkdm9\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.621758 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a72f71c-ec9c-4c47-b888-51c35fbae050-logs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.628891 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.638488 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-config-data\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.641032 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-public-tls-certs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.644865 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.651563 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkdm9\" (UniqueName: \"kubernetes.io/projected/6a72f71c-ec9c-4c47-b888-51c35fbae050-kube-api-access-wkdm9\") pod \"nova-api-0\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " pod="openstack/nova-api-0" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.667453 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-92xms"] Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.669345 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.672564 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.672589 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.699557 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-92xms"] Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.723329 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-config-data\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.723436 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.723567 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-scripts\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.723643 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cx6z\" (UniqueName: \"kubernetes.io/projected/98076d4d-dc13-4a90-9deb-72dc6d9b6095-kube-api-access-7cx6z\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.826291 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cx6z\" (UniqueName: \"kubernetes.io/projected/98076d4d-dc13-4a90-9deb-72dc6d9b6095-kube-api-access-7cx6z\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.827078 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-config-data\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.827151 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.827258 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-scripts\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.839396 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.840408 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-scripts\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.842422 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-config-data\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.884509 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cx6z\" (UniqueName: \"kubernetes.io/projected/98076d4d-dc13-4a90-9deb-72dc6d9b6095-kube-api-access-7cx6z\") pod \"nova-cell1-cell-mapping-92xms\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:26 crc kubenswrapper[4725]: I1202 13:25:26.904035 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:27 crc kubenswrapper[4725]: I1202 13:25:27.057595 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:27 crc kubenswrapper[4725]: I1202 13:25:27.289975 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="548320bf-6a44-424e-af63-6900934d7ae3" path="/var/lib/kubelet/pods/548320bf-6a44-424e-af63-6900934d7ae3/volumes" Dec 02 13:25:27 crc kubenswrapper[4725]: I1202 13:25:27.393420 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0","Type":"ContainerStarted","Data":"788ee9d7f9c680800cc8c7fa6af4dda6ea3a6b21a9b95c84462b6b9077e2447d"} Dec 02 13:25:27 crc kubenswrapper[4725]: I1202 13:25:27.523258 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:27 crc kubenswrapper[4725]: I1202 13:25:27.642410 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-92xms"] Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.405786 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a72f71c-ec9c-4c47-b888-51c35fbae050","Type":"ContainerStarted","Data":"c5e44a0b1ab0250d3acd27708fb2e8ece33ca14d37ac328b836021156ed7ca9b"} Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.406304 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a72f71c-ec9c-4c47-b888-51c35fbae050","Type":"ContainerStarted","Data":"44ed2cad4986e9f3fbaa33193b8751d418779348fd30be9177344d9cc05b684e"} Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.406325 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a72f71c-ec9c-4c47-b888-51c35fbae050","Type":"ContainerStarted","Data":"e5063bbba88779c6753c802e9220a126f4eb8d750074c7d62a34b08f24526d59"} Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.408748 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0","Type":"ContainerStarted","Data":"c07487164356489e217a86b057e1de6430e2fcecc98b793377824ddeed58dfc2"} Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.410796 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92xms" event={"ID":"98076d4d-dc13-4a90-9deb-72dc6d9b6095","Type":"ContainerStarted","Data":"ae0806dcd2c174d69752d508447b56df6d47bbfcb915221751835a7355a1234a"} Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.410848 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92xms" event={"ID":"98076d4d-dc13-4a90-9deb-72dc6d9b6095","Type":"ContainerStarted","Data":"1883b4596fc9ed3033e28a6922fa2489f748640ad6cf9d2032d72cdb40ff3c6d"} Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.437783 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.437764213 podStartE2EDuration="2.437764213s" podCreationTimestamp="2025-12-02 13:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:28.429173217 +0000 UTC m=+1259.385814922" watchObservedRunningTime="2025-12-02 13:25:28.437764213 +0000 UTC m=+1259.394405918" Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.459640 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-92xms" podStartSLOduration=2.459616432 podStartE2EDuration="2.459616432s" podCreationTimestamp="2025-12-02 13:25:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:28.451927689 +0000 UTC m=+1259.408569404" watchObservedRunningTime="2025-12-02 13:25:28.459616432 +0000 UTC m=+1259.416258137" Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.766649 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.848664 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gb8s7"] Dec 02 13:25:28 crc kubenswrapper[4725]: I1202 13:25:28.849109 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="dnsmasq-dns" containerID="cri-o://66c3b969706fc07653030d20d28aeb6ab12b72fbfdd14d4081ecc667d8220507" gracePeriod=10 Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.083978 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.188:5353: connect: connection refused" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.464559 4725 generic.go:334] "Generic (PLEG): container finished" podID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerID="66c3b969706fc07653030d20d28aeb6ab12b72fbfdd14d4081ecc667d8220507" exitCode=0 Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.464637 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" event={"ID":"7c41b73e-5453-4e4f-8f35-dcb324b78e4e","Type":"ContainerDied","Data":"66c3b969706fc07653030d20d28aeb6ab12b72fbfdd14d4081ecc667d8220507"} Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.474233 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.601948 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-nb\") pod \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.602043 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kxwb\" (UniqueName: \"kubernetes.io/projected/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-kube-api-access-5kxwb\") pod \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.602218 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-swift-storage-0\") pod \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.602283 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-config\") pod \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.602331 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-sb\") pod \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.602502 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-svc\") pod \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\" (UID: \"7c41b73e-5453-4e4f-8f35-dcb324b78e4e\") " Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.610684 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-kube-api-access-5kxwb" (OuterVolumeSpecName: "kube-api-access-5kxwb") pod "7c41b73e-5453-4e4f-8f35-dcb324b78e4e" (UID: "7c41b73e-5453-4e4f-8f35-dcb324b78e4e"). InnerVolumeSpecName "kube-api-access-5kxwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.681254 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7c41b73e-5453-4e4f-8f35-dcb324b78e4e" (UID: "7c41b73e-5453-4e4f-8f35-dcb324b78e4e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.691792 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7c41b73e-5453-4e4f-8f35-dcb324b78e4e" (UID: "7c41b73e-5453-4e4f-8f35-dcb324b78e4e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.704038 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-config" (OuterVolumeSpecName: "config") pod "7c41b73e-5453-4e4f-8f35-dcb324b78e4e" (UID: "7c41b73e-5453-4e4f-8f35-dcb324b78e4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.705808 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.705853 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kxwb\" (UniqueName: \"kubernetes.io/projected/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-kube-api-access-5kxwb\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.705875 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.705892 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.705978 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7c41b73e-5453-4e4f-8f35-dcb324b78e4e" (UID: "7c41b73e-5453-4e4f-8f35-dcb324b78e4e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.715039 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c41b73e-5453-4e4f-8f35-dcb324b78e4e" (UID: "7c41b73e-5453-4e4f-8f35-dcb324b78e4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.808069 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:29 crc kubenswrapper[4725]: I1202 13:25:29.808132 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c41b73e-5453-4e4f-8f35-dcb324b78e4e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.477429 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0","Type":"ContainerStarted","Data":"f908d59bd1cbb88ecc263fb7581e33ef7888ad2a5ad974bf544899290a5cad5e"} Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.480148 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" event={"ID":"7c41b73e-5453-4e4f-8f35-dcb324b78e4e","Type":"ContainerDied","Data":"443bb844e045c4ad66122ada9ac01197e4f31f88392968ff97dd59ee3b00ed9f"} Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.480199 4725 scope.go:117] "RemoveContainer" containerID="66c3b969706fc07653030d20d28aeb6ab12b72fbfdd14d4081ecc667d8220507" Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.480273 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gb8s7" Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.510910 4725 scope.go:117] "RemoveContainer" containerID="09e8b28de4ba97b70df54d62e87722401e110d7e731fd53e3379c3e0cb14cbfc" Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.530358 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gb8s7"] Dec 02 13:25:30 crc kubenswrapper[4725]: I1202 13:25:30.544690 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gb8s7"] Dec 02 13:25:31 crc kubenswrapper[4725]: I1202 13:25:31.280415 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" path="/var/lib/kubelet/pods/7c41b73e-5453-4e4f-8f35-dcb324b78e4e/volumes" Dec 02 13:25:33 crc kubenswrapper[4725]: I1202 13:25:33.520620 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e7bb4e-6337-40e2-bea3-ef63e192d2c0","Type":"ContainerStarted","Data":"cd86cd7359492d551bc820505ceb0c2ba20b9a34af01a4f336095cfdb6100e5c"} Dec 02 13:25:33 crc kubenswrapper[4725]: I1202 13:25:33.521323 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 13:25:33 crc kubenswrapper[4725]: I1202 13:25:33.548353 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.416134049 podStartE2EDuration="9.548333893s" podCreationTimestamp="2025-12-02 13:25:24 +0000 UTC" firstStartedPulling="2025-12-02 13:25:25.379438986 +0000 UTC m=+1256.336080681" lastFinishedPulling="2025-12-02 13:25:32.51163884 +0000 UTC m=+1263.468280525" observedRunningTime="2025-12-02 13:25:33.548204579 +0000 UTC m=+1264.504846284" watchObservedRunningTime="2025-12-02 13:25:33.548333893 +0000 UTC m=+1264.504975588" Dec 02 13:25:34 crc kubenswrapper[4725]: I1202 13:25:34.533415 4725 generic.go:334] "Generic (PLEG): container finished" podID="98076d4d-dc13-4a90-9deb-72dc6d9b6095" containerID="ae0806dcd2c174d69752d508447b56df6d47bbfcb915221751835a7355a1234a" exitCode=0 Dec 02 13:25:34 crc kubenswrapper[4725]: I1202 13:25:34.533511 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92xms" event={"ID":"98076d4d-dc13-4a90-9deb-72dc6d9b6095","Type":"ContainerDied","Data":"ae0806dcd2c174d69752d508447b56df6d47bbfcb915221751835a7355a1234a"} Dec 02 13:25:35 crc kubenswrapper[4725]: I1202 13:25:35.962782 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.048289 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cx6z\" (UniqueName: \"kubernetes.io/projected/98076d4d-dc13-4a90-9deb-72dc6d9b6095-kube-api-access-7cx6z\") pod \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.048435 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-scripts\") pod \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.048479 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-combined-ca-bundle\") pod \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.048570 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-config-data\") pod \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\" (UID: \"98076d4d-dc13-4a90-9deb-72dc6d9b6095\") " Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.059783 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-scripts" (OuterVolumeSpecName: "scripts") pod "98076d4d-dc13-4a90-9deb-72dc6d9b6095" (UID: "98076d4d-dc13-4a90-9deb-72dc6d9b6095"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.073407 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98076d4d-dc13-4a90-9deb-72dc6d9b6095-kube-api-access-7cx6z" (OuterVolumeSpecName: "kube-api-access-7cx6z") pod "98076d4d-dc13-4a90-9deb-72dc6d9b6095" (UID: "98076d4d-dc13-4a90-9deb-72dc6d9b6095"). InnerVolumeSpecName "kube-api-access-7cx6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.087310 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-config-data" (OuterVolumeSpecName: "config-data") pod "98076d4d-dc13-4a90-9deb-72dc6d9b6095" (UID: "98076d4d-dc13-4a90-9deb-72dc6d9b6095"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.089735 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98076d4d-dc13-4a90-9deb-72dc6d9b6095" (UID: "98076d4d-dc13-4a90-9deb-72dc6d9b6095"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.151214 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cx6z\" (UniqueName: \"kubernetes.io/projected/98076d4d-dc13-4a90-9deb-72dc6d9b6095-kube-api-access-7cx6z\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.151268 4725 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.151284 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.151302 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98076d4d-dc13-4a90-9deb-72dc6d9b6095-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.557252 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-92xms" event={"ID":"98076d4d-dc13-4a90-9deb-72dc6d9b6095","Type":"ContainerDied","Data":"1883b4596fc9ed3033e28a6922fa2489f748640ad6cf9d2032d72cdb40ff3c6d"} Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.557869 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1883b4596fc9ed3033e28a6922fa2489f748640ad6cf9d2032d72cdb40ff3c6d" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.557418 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-92xms" Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.751658 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.752092 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-log" containerID="cri-o://44ed2cad4986e9f3fbaa33193b8751d418779348fd30be9177344d9cc05b684e" gracePeriod=30 Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.752870 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-api" containerID="cri-o://c5e44a0b1ab0250d3acd27708fb2e8ece33ca14d37ac328b836021156ed7ca9b" gracePeriod=30 Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.763057 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.763440 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" containerName="nova-scheduler-scheduler" containerID="cri-o://15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f" gracePeriod=30 Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.847396 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.847759 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-log" containerID="cri-o://771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5" gracePeriod=30 Dec 02 13:25:36 crc kubenswrapper[4725]: I1202 13:25:36.848011 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-metadata" containerID="cri-o://632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637" gracePeriod=30 Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.574318 4725 generic.go:334] "Generic (PLEG): container finished" podID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerID="771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5" exitCode=143 Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.574498 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"15e51fdb-0e7a-46e1-959f-b384ebfe25bd","Type":"ContainerDied","Data":"771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5"} Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.577658 4725 generic.go:334] "Generic (PLEG): container finished" podID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerID="c5e44a0b1ab0250d3acd27708fb2e8ece33ca14d37ac328b836021156ed7ca9b" exitCode=0 Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.577745 4725 generic.go:334] "Generic (PLEG): container finished" podID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerID="44ed2cad4986e9f3fbaa33193b8751d418779348fd30be9177344d9cc05b684e" exitCode=143 Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.577783 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a72f71c-ec9c-4c47-b888-51c35fbae050","Type":"ContainerDied","Data":"c5e44a0b1ab0250d3acd27708fb2e8ece33ca14d37ac328b836021156ed7ca9b"} Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.577827 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a72f71c-ec9c-4c47-b888-51c35fbae050","Type":"ContainerDied","Data":"44ed2cad4986e9f3fbaa33193b8751d418779348fd30be9177344d9cc05b684e"} Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.577843 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6a72f71c-ec9c-4c47-b888-51c35fbae050","Type":"ContainerDied","Data":"e5063bbba88779c6753c802e9220a126f4eb8d750074c7d62a34b08f24526d59"} Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.577856 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5063bbba88779c6753c802e9220a126f4eb8d750074c7d62a34b08f24526d59" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.584340 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.691211 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a72f71c-ec9c-4c47-b888-51c35fbae050-logs\") pod \"6a72f71c-ec9c-4c47-b888-51c35fbae050\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.691356 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkdm9\" (UniqueName: \"kubernetes.io/projected/6a72f71c-ec9c-4c47-b888-51c35fbae050-kube-api-access-wkdm9\") pod \"6a72f71c-ec9c-4c47-b888-51c35fbae050\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.691449 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-config-data\") pod \"6a72f71c-ec9c-4c47-b888-51c35fbae050\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.691602 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-combined-ca-bundle\") pod \"6a72f71c-ec9c-4c47-b888-51c35fbae050\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.691674 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-internal-tls-certs\") pod \"6a72f71c-ec9c-4c47-b888-51c35fbae050\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.691768 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-public-tls-certs\") pod \"6a72f71c-ec9c-4c47-b888-51c35fbae050\" (UID: \"6a72f71c-ec9c-4c47-b888-51c35fbae050\") " Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.692419 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a72f71c-ec9c-4c47-b888-51c35fbae050-logs" (OuterVolumeSpecName: "logs") pod "6a72f71c-ec9c-4c47-b888-51c35fbae050" (UID: "6a72f71c-ec9c-4c47-b888-51c35fbae050"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.700650 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a72f71c-ec9c-4c47-b888-51c35fbae050-kube-api-access-wkdm9" (OuterVolumeSpecName: "kube-api-access-wkdm9") pod "6a72f71c-ec9c-4c47-b888-51c35fbae050" (UID: "6a72f71c-ec9c-4c47-b888-51c35fbae050"). InnerVolumeSpecName "kube-api-access-wkdm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.730215 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a72f71c-ec9c-4c47-b888-51c35fbae050" (UID: "6a72f71c-ec9c-4c47-b888-51c35fbae050"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.741406 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-config-data" (OuterVolumeSpecName: "config-data") pod "6a72f71c-ec9c-4c47-b888-51c35fbae050" (UID: "6a72f71c-ec9c-4c47-b888-51c35fbae050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.761727 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6a72f71c-ec9c-4c47-b888-51c35fbae050" (UID: "6a72f71c-ec9c-4c47-b888-51c35fbae050"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.775325 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6a72f71c-ec9c-4c47-b888-51c35fbae050" (UID: "6a72f71c-ec9c-4c47-b888-51c35fbae050"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.795294 4725 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.795342 4725 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.795356 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a72f71c-ec9c-4c47-b888-51c35fbae050-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.795369 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkdm9\" (UniqueName: \"kubernetes.io/projected/6a72f71c-ec9c-4c47-b888-51c35fbae050-kube-api-access-wkdm9\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.795387 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:37 crc kubenswrapper[4725]: I1202 13:25:37.795401 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a72f71c-ec9c-4c47-b888-51c35fbae050-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.592483 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.642381 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.665846 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.683994 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:38 crc kubenswrapper[4725]: E1202 13:25:38.684547 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="init" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684566 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="init" Dec 02 13:25:38 crc kubenswrapper[4725]: E1202 13:25:38.684581 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="dnsmasq-dns" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684588 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="dnsmasq-dns" Dec 02 13:25:38 crc kubenswrapper[4725]: E1202 13:25:38.684604 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-api" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684612 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-api" Dec 02 13:25:38 crc kubenswrapper[4725]: E1202 13:25:38.684626 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98076d4d-dc13-4a90-9deb-72dc6d9b6095" containerName="nova-manage" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684633 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="98076d4d-dc13-4a90-9deb-72dc6d9b6095" containerName="nova-manage" Dec 02 13:25:38 crc kubenswrapper[4725]: E1202 13:25:38.684659 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-log" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684666 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-log" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684885 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c41b73e-5453-4e4f-8f35-dcb324b78e4e" containerName="dnsmasq-dns" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684914 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-log" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684927 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="98076d4d-dc13-4a90-9deb-72dc6d9b6095" containerName="nova-manage" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.684943 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" containerName="nova-api-api" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.686125 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.688625 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.689073 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.690296 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.696441 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.717740 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa1cff6-4533-4ba6-aeac-86a813cf8516-logs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.717853 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.717923 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.717983 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-config-data\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.718035 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wklgk\" (UniqueName: \"kubernetes.io/projected/3aa1cff6-4533-4ba6-aeac-86a813cf8516-kube-api-access-wklgk\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.718085 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-public-tls-certs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.822522 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-config-data\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.822648 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wklgk\" (UniqueName: \"kubernetes.io/projected/3aa1cff6-4533-4ba6-aeac-86a813cf8516-kube-api-access-wklgk\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.822770 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-public-tls-certs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.822937 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa1cff6-4533-4ba6-aeac-86a813cf8516-logs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.823046 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.823121 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.825070 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3aa1cff6-4533-4ba6-aeac-86a813cf8516-logs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.830100 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-public-tls-certs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.830241 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.831964 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-config-data\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.836223 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa1cff6-4533-4ba6-aeac-86a813cf8516-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:38 crc kubenswrapper[4725]: I1202 13:25:38.852708 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wklgk\" (UniqueName: \"kubernetes.io/projected/3aa1cff6-4533-4ba6-aeac-86a813cf8516-kube-api-access-wklgk\") pod \"nova-api-0\" (UID: \"3aa1cff6-4533-4ba6-aeac-86a813cf8516\") " pod="openstack/nova-api-0" Dec 02 13:25:39 crc kubenswrapper[4725]: I1202 13:25:39.015214 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 13:25:39 crc kubenswrapper[4725]: I1202 13:25:39.287963 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a72f71c-ec9c-4c47-b888-51c35fbae050" path="/var/lib/kubelet/pods/6a72f71c-ec9c-4c47-b888-51c35fbae050/volumes" Dec 02 13:25:39 crc kubenswrapper[4725]: I1202 13:25:39.498303 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 13:25:39 crc kubenswrapper[4725]: W1202 13:25:39.501418 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aa1cff6_4533_4ba6_aeac_86a813cf8516.slice/crio-67d280c54633ed74bf07886716cb736e988cd08adbedeef75ef3f36b362072c0 WatchSource:0}: Error finding container 67d280c54633ed74bf07886716cb736e988cd08adbedeef75ef3f36b362072c0: Status 404 returned error can't find the container with id 67d280c54633ed74bf07886716cb736e988cd08adbedeef75ef3f36b362072c0 Dec 02 13:25:39 crc kubenswrapper[4725]: I1202 13:25:39.607323 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3aa1cff6-4533-4ba6-aeac-86a813cf8516","Type":"ContainerStarted","Data":"67d280c54633ed74bf07886716cb736e988cd08adbedeef75ef3f36b362072c0"} Dec 02 13:25:39 crc kubenswrapper[4725]: I1202 13:25:39.993759 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:51906->10.217.0.194:8775: read: connection reset by peer" Dec 02 13:25:39 crc kubenswrapper[4725]: I1202 13:25:39.993781 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:51912->10.217.0.194:8775: read: connection reset by peer" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.507348 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.625266 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3aa1cff6-4533-4ba6-aeac-86a813cf8516","Type":"ContainerStarted","Data":"3a0e31d5bd3fa28bc7bd2c8fb109f0ce24f83a74ffdc3efd3c5bbc51a566166e"} Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.625342 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3aa1cff6-4533-4ba6-aeac-86a813cf8516","Type":"ContainerStarted","Data":"0f8e5506a506ba8d57526ec24bc7c904d27dbb44c0a18209e3a713baa7bd6a65"} Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.629011 4725 generic.go:334] "Generic (PLEG): container finished" podID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerID="632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637" exitCode=0 Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.629073 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"15e51fdb-0e7a-46e1-959f-b384ebfe25bd","Type":"ContainerDied","Data":"632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637"} Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.629114 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"15e51fdb-0e7a-46e1-959f-b384ebfe25bd","Type":"ContainerDied","Data":"3a491a180a2751841c4ac8caf628414c627ec52118ae6542511ab7d1622bdaa7"} Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.629137 4725 scope.go:117] "RemoveContainer" containerID="632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.629366 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.657524 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.657500177 podStartE2EDuration="2.657500177s" podCreationTimestamp="2025-12-02 13:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:40.644049169 +0000 UTC m=+1271.600690864" watchObservedRunningTime="2025-12-02 13:25:40.657500177 +0000 UTC m=+1271.614141872" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.666864 4725 scope.go:117] "RemoveContainer" containerID="771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.670553 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-config-data\") pod \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.670703 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-combined-ca-bundle\") pod \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.670748 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjhl8\" (UniqueName: \"kubernetes.io/projected/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-kube-api-access-pjhl8\") pod \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.670782 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-nova-metadata-tls-certs\") pod \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.670836 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-logs\") pod \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\" (UID: \"15e51fdb-0e7a-46e1-959f-b384ebfe25bd\") " Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.672056 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-logs" (OuterVolumeSpecName: "logs") pod "15e51fdb-0e7a-46e1-959f-b384ebfe25bd" (UID: "15e51fdb-0e7a-46e1-959f-b384ebfe25bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.677601 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-kube-api-access-pjhl8" (OuterVolumeSpecName: "kube-api-access-pjhl8") pod "15e51fdb-0e7a-46e1-959f-b384ebfe25bd" (UID: "15e51fdb-0e7a-46e1-959f-b384ebfe25bd"). InnerVolumeSpecName "kube-api-access-pjhl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.702385 4725 scope.go:117] "RemoveContainer" containerID="632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637" Dec 02 13:25:40 crc kubenswrapper[4725]: E1202 13:25:40.703311 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637\": container with ID starting with 632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637 not found: ID does not exist" containerID="632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.703429 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637"} err="failed to get container status \"632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637\": rpc error: code = NotFound desc = could not find container \"632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637\": container with ID starting with 632ba90b98e69f2c69b47818b39350cfbd5addb88d2e54b26536291c2e4cc637 not found: ID does not exist" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.703586 4725 scope.go:117] "RemoveContainer" containerID="771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5" Dec 02 13:25:40 crc kubenswrapper[4725]: E1202 13:25:40.705622 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5\": container with ID starting with 771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5 not found: ID does not exist" containerID="771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.705705 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5"} err="failed to get container status \"771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5\": rpc error: code = NotFound desc = could not find container \"771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5\": container with ID starting with 771c07134124075d3fff6de24699e47c24f5051c4e4caea0795176f57c45e7a5 not found: ID does not exist" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.706167 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-config-data" (OuterVolumeSpecName: "config-data") pod "15e51fdb-0e7a-46e1-959f-b384ebfe25bd" (UID: "15e51fdb-0e7a-46e1-959f-b384ebfe25bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.724888 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15e51fdb-0e7a-46e1-959f-b384ebfe25bd" (UID: "15e51fdb-0e7a-46e1-959f-b384ebfe25bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.752970 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "15e51fdb-0e7a-46e1-959f-b384ebfe25bd" (UID: "15e51fdb-0e7a-46e1-959f-b384ebfe25bd"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.776701 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjhl8\" (UniqueName: \"kubernetes.io/projected/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-kube-api-access-pjhl8\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.776745 4725 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.776759 4725 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-logs\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.776796 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.776810 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15e51fdb-0e7a-46e1-959f-b384ebfe25bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:40 crc kubenswrapper[4725]: I1202 13:25:40.984171 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.000570 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.024583 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.040897 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-metadata" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.040928 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-metadata" Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.040970 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-log" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.040979 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-log" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.041454 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-log" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.041527 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" containerName="nova-metadata-metadata" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.047371 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.061273 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.062553 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.062766 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.085011 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-logs\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.085067 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.085168 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-config-data\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.085321 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzrj6\" (UniqueName: \"kubernetes.io/projected/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-kube-api-access-bzrj6\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.085343 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.152654 4725 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15e51fdb_0e7a_46e1_959f_b384ebfe25bd.slice/crio-3a491a180a2751841c4ac8caf628414c627ec52118ae6542511ab7d1622bdaa7\": RecentStats: unable to find data in memory cache]" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.186688 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzrj6\" (UniqueName: \"kubernetes.io/projected/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-kube-api-access-bzrj6\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.186733 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.186805 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-logs\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.186828 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.186881 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-config-data\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.190405 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-logs\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.197015 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.197072 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-config-data\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.197096 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.206990 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzrj6\" (UniqueName: \"kubernetes.io/projected/5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d-kube-api-access-bzrj6\") pod \"nova-metadata-0\" (UID: \"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d\") " pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.258730 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.286672 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15e51fdb-0e7a-46e1-959f-b384ebfe25bd" path="/var/lib/kubelet/pods/15e51fdb-0e7a-46e1-959f-b384ebfe25bd/volumes" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.650821 4725 generic.go:334] "Generic (PLEG): container finished" podID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" containerID="15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f" exitCode=0 Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.650898 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88afcfe6-5dc2-4ffc-a79b-4c46585122bf","Type":"ContainerDied","Data":"15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f"} Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.656161 4725 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f is running failed: container process not found" containerID="15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.656827 4725 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f is running failed: container process not found" containerID="15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.657190 4725 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f is running failed: container process not found" containerID="15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 13:25:41 crc kubenswrapper[4725]: E1202 13:25:41.657255 4725 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" containerName="nova-scheduler-scheduler" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.804404 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.901051 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-combined-ca-bundle\") pod \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.901171 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7fxh\" (UniqueName: \"kubernetes.io/projected/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-kube-api-access-v7fxh\") pod \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.901215 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-config-data\") pod \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\" (UID: \"88afcfe6-5dc2-4ffc-a79b-4c46585122bf\") " Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.907712 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-kube-api-access-v7fxh" (OuterVolumeSpecName: "kube-api-access-v7fxh") pod "88afcfe6-5dc2-4ffc-a79b-4c46585122bf" (UID: "88afcfe6-5dc2-4ffc-a79b-4c46585122bf"). InnerVolumeSpecName "kube-api-access-v7fxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.933645 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-config-data" (OuterVolumeSpecName: "config-data") pod "88afcfe6-5dc2-4ffc-a79b-4c46585122bf" (UID: "88afcfe6-5dc2-4ffc-a79b-4c46585122bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.939681 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88afcfe6-5dc2-4ffc-a79b-4c46585122bf" (UID: "88afcfe6-5dc2-4ffc-a79b-4c46585122bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:25:41 crc kubenswrapper[4725]: I1202 13:25:41.942059 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 13:25:41 crc kubenswrapper[4725]: W1202 13:25:41.957892 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b6a3cb0_e50a_4c5b_a83f_87208b3fe04d.slice/crio-019740465bbcb3b6d94c644d9c94260643624b9ba5c90101b69f71396d1c6732 WatchSource:0}: Error finding container 019740465bbcb3b6d94c644d9c94260643624b9ba5c90101b69f71396d1c6732: Status 404 returned error can't find the container with id 019740465bbcb3b6d94c644d9c94260643624b9ba5c90101b69f71396d1c6732 Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.003735 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.003779 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7fxh\" (UniqueName: \"kubernetes.io/projected/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-kube-api-access-v7fxh\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.003796 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88afcfe6-5dc2-4ffc-a79b-4c46585122bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.660827 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d","Type":"ContainerStarted","Data":"4de13ff1cd0a323618fbdc4b808b9eb61a1935a9ffd68275b78a4ee78a630d1c"} Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.661136 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d","Type":"ContainerStarted","Data":"88df37ef4550b3da5f155f26a00eb52b5ebf4fe49ade0dd52c75c0211b379b3c"} Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.661153 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d","Type":"ContainerStarted","Data":"019740465bbcb3b6d94c644d9c94260643624b9ba5c90101b69f71396d1c6732"} Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.663510 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"88afcfe6-5dc2-4ffc-a79b-4c46585122bf","Type":"ContainerDied","Data":"04ebd6aa534134bb21715c3d8c6434a5835e141c4d2544487acfd854d5ebd02d"} Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.663565 4725 scope.go:117] "RemoveContainer" containerID="15d619dc58b03aa32aa87902a5db0ed626e29f60822c50f8ce57fb24ed91739f" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.663578 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.694344 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.694327324 podStartE2EDuration="2.694327324s" podCreationTimestamp="2025-12-02 13:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:42.692201171 +0000 UTC m=+1273.648842876" watchObservedRunningTime="2025-12-02 13:25:42.694327324 +0000 UTC m=+1273.650969009" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.731531 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.740755 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.758609 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:25:42 crc kubenswrapper[4725]: E1202 13:25:42.759018 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" containerName="nova-scheduler-scheduler" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.759034 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" containerName="nova-scheduler-scheduler" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.759251 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" containerName="nova-scheduler-scheduler" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.759871 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.765753 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.823655 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.920973 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl2ss\" (UniqueName: \"kubernetes.io/projected/c1a212a8-1993-4689-988a-b6fc12ad4cb3-kube-api-access-xl2ss\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.921037 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a212a8-1993-4689-988a-b6fc12ad4cb3-config-data\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:42 crc kubenswrapper[4725]: I1202 13:25:42.921057 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a212a8-1993-4689-988a-b6fc12ad4cb3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.022900 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl2ss\" (UniqueName: \"kubernetes.io/projected/c1a212a8-1993-4689-988a-b6fc12ad4cb3-kube-api-access-xl2ss\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.022959 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a212a8-1993-4689-988a-b6fc12ad4cb3-config-data\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.022983 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a212a8-1993-4689-988a-b6fc12ad4cb3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.027910 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a212a8-1993-4689-988a-b6fc12ad4cb3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.028317 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1a212a8-1993-4689-988a-b6fc12ad4cb3-config-data\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.041175 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl2ss\" (UniqueName: \"kubernetes.io/projected/c1a212a8-1993-4689-988a-b6fc12ad4cb3-kube-api-access-xl2ss\") pod \"nova-scheduler-0\" (UID: \"c1a212a8-1993-4689-988a-b6fc12ad4cb3\") " pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.086657 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.281826 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88afcfe6-5dc2-4ffc-a79b-4c46585122bf" path="/var/lib/kubelet/pods/88afcfe6-5dc2-4ffc-a79b-4c46585122bf/volumes" Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.531334 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 13:25:43 crc kubenswrapper[4725]: I1202 13:25:43.674534 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c1a212a8-1993-4689-988a-b6fc12ad4cb3","Type":"ContainerStarted","Data":"5cd08d66d1eaf3f902fae98ffad8d6673316ccc4ec249aeee278bf2824719694"} Dec 02 13:25:44 crc kubenswrapper[4725]: I1202 13:25:44.686670 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c1a212a8-1993-4689-988a-b6fc12ad4cb3","Type":"ContainerStarted","Data":"205a56e22f3d237bd48714194911d1ffb3443b7414c18d96b3ad8275c8bb5b15"} Dec 02 13:25:44 crc kubenswrapper[4725]: I1202 13:25:44.718271 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.718249835 podStartE2EDuration="2.718249835s" podCreationTimestamp="2025-12-02 13:25:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:25:44.707022423 +0000 UTC m=+1275.663664118" watchObservedRunningTime="2025-12-02 13:25:44.718249835 +0000 UTC m=+1275.674891530" Dec 02 13:25:46 crc kubenswrapper[4725]: I1202 13:25:46.259391 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:25:46 crc kubenswrapper[4725]: I1202 13:25:46.259435 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 13:25:48 crc kubenswrapper[4725]: I1202 13:25:48.087295 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 13:25:49 crc kubenswrapper[4725]: I1202 13:25:49.015863 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 13:25:49 crc kubenswrapper[4725]: I1202 13:25:49.016204 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 13:25:50 crc kubenswrapper[4725]: I1202 13:25:50.029772 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3aa1cff6-4533-4ba6-aeac-86a813cf8516" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:50 crc kubenswrapper[4725]: I1202 13:25:50.029781 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3aa1cff6-4533-4ba6-aeac-86a813cf8516" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:51 crc kubenswrapper[4725]: I1202 13:25:51.259171 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 13:25:51 crc kubenswrapper[4725]: I1202 13:25:51.259216 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 13:25:52 crc kubenswrapper[4725]: I1202 13:25:52.049811 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="53e3272a-6a56-4bba-8a07-48997cfc1dc8" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.192:3000/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:52 crc kubenswrapper[4725]: I1202 13:25:52.273634 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:52 crc kubenswrapper[4725]: I1202 13:25:52.273634 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 13:25:53 crc kubenswrapper[4725]: I1202 13:25:53.087343 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 13:25:53 crc kubenswrapper[4725]: I1202 13:25:53.133517 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 13:25:53 crc kubenswrapper[4725]: I1202 13:25:53.804650 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 13:25:54 crc kubenswrapper[4725]: I1202 13:25:54.284938 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:25:54 crc kubenswrapper[4725]: I1202 13:25:54.285002 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:25:54 crc kubenswrapper[4725]: I1202 13:25:54.779912 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 02 13:25:59 crc kubenswrapper[4725]: I1202 13:25:59.022483 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 13:25:59 crc kubenswrapper[4725]: I1202 13:25:59.023174 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 13:25:59 crc kubenswrapper[4725]: I1202 13:25:59.023208 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 13:25:59 crc kubenswrapper[4725]: I1202 13:25:59.029027 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 13:25:59 crc kubenswrapper[4725]: I1202 13:25:59.830644 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 13:25:59 crc kubenswrapper[4725]: I1202 13:25:59.835531 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 13:26:01 crc kubenswrapper[4725]: I1202 13:26:01.265192 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 13:26:01 crc kubenswrapper[4725]: I1202 13:26:01.265976 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 13:26:01 crc kubenswrapper[4725]: I1202 13:26:01.278839 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 13:26:01 crc kubenswrapper[4725]: I1202 13:26:01.858046 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 13:26:10 crc kubenswrapper[4725]: I1202 13:26:10.275181 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:26:11 crc kubenswrapper[4725]: I1202 13:26:11.328621 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:26:14 crc kubenswrapper[4725]: I1202 13:26:14.826499 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="rabbitmq" containerID="cri-o://4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f" gracePeriod=604796 Dec 02 13:26:15 crc kubenswrapper[4725]: I1202 13:26:15.672648 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="rabbitmq" containerID="cri-o://375c371f1dd02d3aa09718b311330c557f3512342b53e32ba612467bf474fbdc" gracePeriod=604796 Dec 02 13:26:18 crc kubenswrapper[4725]: I1202 13:26:18.611193 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 02 13:26:18 crc kubenswrapper[4725]: I1202 13:26:18.929079 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.399506 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.484879 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.484971 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-erlang-cookie\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485075 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-confd\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485142 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-tls\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485187 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d630b03-f4dd-48ac-a72b-b457e3651c76-pod-info\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485231 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-server-conf\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485268 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-plugins\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485298 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d630b03-f4dd-48ac-a72b-b457e3651c76-erlang-cookie-secret\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485371 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-config-data\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485487 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cln48\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-kube-api-access-cln48\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.485561 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-plugins-conf\") pod \"0d630b03-f4dd-48ac-a72b-b457e3651c76\" (UID: \"0d630b03-f4dd-48ac-a72b-b457e3651c76\") " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.486000 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.486134 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.486869 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.488931 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.501927 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.501926 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d630b03-f4dd-48ac-a72b-b457e3651c76-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.502103 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.503414 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0d630b03-f4dd-48ac-a72b-b457e3651c76-pod-info" (OuterVolumeSpecName: "pod-info") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.507667 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-kube-api-access-cln48" (OuterVolumeSpecName: "kube-api-access-cln48") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "kube-api-access-cln48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.549510 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-config-data" (OuterVolumeSpecName: "config-data") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.565488 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-server-conf" (OuterVolumeSpecName: "server-conf") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588317 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588348 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588359 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588368 4725 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0d630b03-f4dd-48ac-a72b-b457e3651c76-pod-info\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588377 4725 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-server-conf\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588387 4725 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0d630b03-f4dd-48ac-a72b-b457e3651c76-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588412 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588421 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cln48\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-kube-api-access-cln48\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.588430 4725 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0d630b03-f4dd-48ac-a72b-b457e3651c76-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.621342 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.664316 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0d630b03-f4dd-48ac-a72b-b457e3651c76" (UID: "0d630b03-f4dd-48ac-a72b-b457e3651c76"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.690729 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:21 crc kubenswrapper[4725]: I1202 13:26:21.691038 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0d630b03-f4dd-48ac-a72b-b457e3651c76-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.038744 4725 generic.go:334] "Generic (PLEG): container finished" podID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerID="4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f" exitCode=0 Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.038811 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d630b03-f4dd-48ac-a72b-b457e3651c76","Type":"ContainerDied","Data":"4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f"} Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.038817 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.038905 4725 scope.go:117] "RemoveContainer" containerID="4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.038889 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0d630b03-f4dd-48ac-a72b-b457e3651c76","Type":"ContainerDied","Data":"f0fffedbd155c725d59b39b63b9c13da0a5b9651c2b1e461e1d4a301a2b91d94"} Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.041288 4725 generic.go:334] "Generic (PLEG): container finished" podID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerID="375c371f1dd02d3aa09718b311330c557f3512342b53e32ba612467bf474fbdc" exitCode=0 Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.041323 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ede4bdcc-2588-4e91-80f7-1bcf0467e861","Type":"ContainerDied","Data":"375c371f1dd02d3aa09718b311330c557f3512342b53e32ba612467bf474fbdc"} Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.083635 4725 scope.go:117] "RemoveContainer" containerID="988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.138935 4725 scope.go:117] "RemoveContainer" containerID="4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.141726 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.159910 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:26:22 crc kubenswrapper[4725]: E1202 13:26:22.160103 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f\": container with ID starting with 4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f not found: ID does not exist" containerID="4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.160155 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f"} err="failed to get container status \"4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f\": rpc error: code = NotFound desc = could not find container \"4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f\": container with ID starting with 4d54f09a5a3bb31beac97b8453d8bb19db195955f8d286492ffc9c4223488e8f not found: ID does not exist" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.160186 4725 scope.go:117] "RemoveContainer" containerID="988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.170125 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:26:22 crc kubenswrapper[4725]: E1202 13:26:22.170853 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113\": container with ID starting with 988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113 not found: ID does not exist" containerID="988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.170888 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113"} err="failed to get container status \"988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113\": rpc error: code = NotFound desc = could not find container \"988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113\": container with ID starting with 988ff3374c8bf333629927534f0187701876a265a9b5713d9b1f414d3df74113 not found: ID does not exist" Dec 02 13:26:22 crc kubenswrapper[4725]: E1202 13:26:22.170995 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="rabbitmq" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.171011 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="rabbitmq" Dec 02 13:26:22 crc kubenswrapper[4725]: E1202 13:26:22.171050 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="setup-container" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.171057 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="setup-container" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.171273 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" containerName="rabbitmq" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.172574 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.174901 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.175730 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.175926 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.176193 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.176413 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-hh5j8" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.176811 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.176947 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.178568 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.300932 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.300978 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301158 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301362 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301418 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301614 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301681 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-config-data\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301750 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301867 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301900 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.301921 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngckr\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-kube-api-access-ngckr\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.316069 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: E1202 13:26:22.358319 4725 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d630b03_f4dd_48ac_a72b_b457e3651c76.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d630b03_f4dd_48ac_a72b_b457e3651c76.slice/crio-f0fffedbd155c725d59b39b63b9c13da0a5b9651c2b1e461e1d4a301a2b91d94\": RecentStats: unable to find data in memory cache]" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.402996 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-plugins\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403092 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4c2m\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-kube-api-access-c4c2m\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403184 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-confd\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403202 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-tls\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403286 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-erlang-cookie\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403325 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ede4bdcc-2588-4e91-80f7-1bcf0467e861-pod-info\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403381 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ede4bdcc-2588-4e91-80f7-1bcf0467e861-erlang-cookie-secret\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403420 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-server-conf\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403476 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-config-data\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403514 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-plugins-conf\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403553 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\" (UID: \"ede4bdcc-2588-4e91-80f7-1bcf0467e861\") " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404126 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-config-data\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.403755 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404166 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404219 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404240 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngckr\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-kube-api-access-ngckr\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404359 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404378 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404396 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404552 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404616 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404709 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.404776 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.406295 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.409118 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.410367 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.411324 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-kube-api-access-c4c2m" (OuterVolumeSpecName: "kube-api-access-c4c2m") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "kube-api-access-c4c2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.411783 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-config-data\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.411870 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.412508 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.413158 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ede4bdcc-2588-4e91-80f7-1bcf0467e861-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.413198 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.413595 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.416507 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.419971 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.421208 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.421662 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.424267 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.430083 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ede4bdcc-2588-4e91-80f7-1bcf0467e861-pod-info" (OuterVolumeSpecName: "pod-info") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.435060 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngckr\" (UniqueName: \"kubernetes.io/projected/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-kube-api-access-ngckr\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.456550 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.457056 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-config-data" (OuterVolumeSpecName: "config-data") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.472139 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7\") " pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.496950 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-server-conf" (OuterVolumeSpecName: "server-conf") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.506897 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.506960 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.506972 4725 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ede4bdcc-2588-4e91-80f7-1bcf0467e861-pod-info\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.506981 4725 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ede4bdcc-2588-4e91-80f7-1bcf0467e861-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.506989 4725 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-server-conf\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.506999 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.507007 4725 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ede4bdcc-2588-4e91-80f7-1bcf0467e861-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.507053 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.507063 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4c2m\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-kube-api-access-c4c2m\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.533714 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.574828 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ede4bdcc-2588-4e91-80f7-1bcf0467e861" (UID: "ede4bdcc-2588-4e91-80f7-1bcf0467e861"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.603715 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.616111 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:22 crc kubenswrapper[4725]: I1202 13:26:22.616151 4725 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ede4bdcc-2588-4e91-80f7-1bcf0467e861-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.057472 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ede4bdcc-2588-4e91-80f7-1bcf0467e861","Type":"ContainerDied","Data":"50642c4080773e0a09deda9800dcf1fb03d37093730302089399718e960cd355"} Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.057826 4725 scope.go:117] "RemoveContainer" containerID="375c371f1dd02d3aa09718b311330c557f3512342b53e32ba612467bf474fbdc" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.057494 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.102035 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.195376 4725 scope.go:117] "RemoveContainer" containerID="6e3ec9fe8664b961c59b197cc682fd6ee84f663d3385bbcb3233ec98fc495c03" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.239913 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.255908 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.279300 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d630b03-f4dd-48ac-a72b-b457e3651c76" path="/var/lib/kubelet/pods/0d630b03-f4dd-48ac-a72b-b457e3651c76/volumes" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.280484 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" path="/var/lib/kubelet/pods/ede4bdcc-2588-4e91-80f7-1bcf0467e861/volumes" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.281841 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:26:23 crc kubenswrapper[4725]: E1202 13:26:23.282255 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="rabbitmq" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.282273 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="rabbitmq" Dec 02 13:26:23 crc kubenswrapper[4725]: E1202 13:26:23.282285 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="setup-container" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.282294 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="setup-container" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.282623 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ede4bdcc-2588-4e91-80f7-1bcf0467e861" containerName="rabbitmq" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.288893 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.289017 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.290755 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.290955 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.291321 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.291552 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.291636 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.291856 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.291868 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-n5vfx" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.430567 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.430689 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.430885 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.430995 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431030 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1774e612-2433-4536-a07c-ef5ac77f2036-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431074 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1774e612-2433-4536-a07c-ef5ac77f2036-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431146 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431262 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431286 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431338 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.431436 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdwsj\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-kube-api-access-fdwsj\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533427 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533585 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdwsj\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-kube-api-access-fdwsj\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533634 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533679 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533722 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533745 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533761 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1774e612-2433-4536-a07c-ef5ac77f2036-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533789 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1774e612-2433-4536-a07c-ef5ac77f2036-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533805 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533855 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.533870 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.534154 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.534243 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.534787 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.535194 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.535266 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.535656 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1774e612-2433-4536-a07c-ef5ac77f2036-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.539114 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1774e612-2433-4536-a07c-ef5ac77f2036-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.539217 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.539574 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1774e612-2433-4536-a07c-ef5ac77f2036-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.539743 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.554000 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdwsj\" (UniqueName: \"kubernetes.io/projected/1774e612-2433-4536-a07c-ef5ac77f2036-kube-api-access-fdwsj\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.563334 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1774e612-2433-4536-a07c-ef5ac77f2036\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:23 crc kubenswrapper[4725]: I1202 13:26:23.645054 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.099574 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7","Type":"ContainerStarted","Data":"179097a711a68f75bb0aa9651e8d79ab8d6bdf20ce7e0600f6d4e0945e19819a"} Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.128921 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 13:26:24 crc kubenswrapper[4725]: W1202 13:26:24.195185 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1774e612_2433_4536_a07c_ef5ac77f2036.slice/crio-e594be8aed9c96acf4de7547ba3864e9d3050c21ccef749d6794d46efe5bab6c WatchSource:0}: Error finding container e594be8aed9c96acf4de7547ba3864e9d3050c21ccef749d6794d46efe5bab6c: Status 404 returned error can't find the container with id e594be8aed9c96acf4de7547ba3864e9d3050c21ccef749d6794d46efe5bab6c Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.284379 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.284714 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.540971 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h57r8"] Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.543072 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.549864 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.552207 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h57r8"] Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652249 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652352 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljr5v\" (UniqueName: \"kubernetes.io/projected/ef8dac21-8448-49a5-bc91-506b522da6d4-kube-api-access-ljr5v\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652422 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652514 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652566 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-config\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652637 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.652706 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.754758 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-config\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.754818 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.754868 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.754908 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.754978 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljr5v\" (UniqueName: \"kubernetes.io/projected/ef8dac21-8448-49a5-bc91-506b522da6d4-kube-api-access-ljr5v\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.755020 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.755107 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.756002 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-config\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.756085 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.756151 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.756572 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.756709 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.756791 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.790511 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljr5v\" (UniqueName: \"kubernetes.io/projected/ef8dac21-8448-49a5-bc91-506b522da6d4-kube-api-access-ljr5v\") pod \"dnsmasq-dns-79bd4cc8c9-h57r8\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:24 crc kubenswrapper[4725]: I1202 13:26:24.863265 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:25 crc kubenswrapper[4725]: I1202 13:26:25.141703 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1774e612-2433-4536-a07c-ef5ac77f2036","Type":"ContainerStarted","Data":"e594be8aed9c96acf4de7547ba3864e9d3050c21ccef749d6794d46efe5bab6c"} Dec 02 13:26:25 crc kubenswrapper[4725]: I1202 13:26:25.196801 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h57r8"] Dec 02 13:26:25 crc kubenswrapper[4725]: W1202 13:26:25.202614 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef8dac21_8448_49a5_bc91_506b522da6d4.slice/crio-5d486096983b3a996aa0904fc79e6c3717ce80efff50dd7419713f7d696c1502 WatchSource:0}: Error finding container 5d486096983b3a996aa0904fc79e6c3717ce80efff50dd7419713f7d696c1502: Status 404 returned error can't find the container with id 5d486096983b3a996aa0904fc79e6c3717ce80efff50dd7419713f7d696c1502 Dec 02 13:26:26 crc kubenswrapper[4725]: I1202 13:26:26.152474 4725 generic.go:334] "Generic (PLEG): container finished" podID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerID="126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092" exitCode=0 Dec 02 13:26:26 crc kubenswrapper[4725]: I1202 13:26:26.152507 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" event={"ID":"ef8dac21-8448-49a5-bc91-506b522da6d4","Type":"ContainerDied","Data":"126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092"} Dec 02 13:26:26 crc kubenswrapper[4725]: I1202 13:26:26.152890 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" event={"ID":"ef8dac21-8448-49a5-bc91-506b522da6d4","Type":"ContainerStarted","Data":"5d486096983b3a996aa0904fc79e6c3717ce80efff50dd7419713f7d696c1502"} Dec 02 13:26:26 crc kubenswrapper[4725]: I1202 13:26:26.155105 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7","Type":"ContainerStarted","Data":"02fe3eb0b73623579b7dc3d091d78fdcb9cf2f7ae0dc68a92b1bc6a7cf0bd957"} Dec 02 13:26:26 crc kubenswrapper[4725]: I1202 13:26:26.157663 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1774e612-2433-4536-a07c-ef5ac77f2036","Type":"ContainerStarted","Data":"495a729d20fb9b3efe34bd15ac2696d0283994fd8f04dbba1802bd4c797cd743"} Dec 02 13:26:27 crc kubenswrapper[4725]: I1202 13:26:27.167902 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" event={"ID":"ef8dac21-8448-49a5-bc91-506b522da6d4","Type":"ContainerStarted","Data":"a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c"} Dec 02 13:26:27 crc kubenswrapper[4725]: I1202 13:26:27.187113 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" podStartSLOduration=3.187091914 podStartE2EDuration="3.187091914s" podCreationTimestamp="2025-12-02 13:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:26:27.182894058 +0000 UTC m=+1318.139535783" watchObservedRunningTime="2025-12-02 13:26:27.187091914 +0000 UTC m=+1318.143733609" Dec 02 13:26:28 crc kubenswrapper[4725]: I1202 13:26:28.176206 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:34 crc kubenswrapper[4725]: I1202 13:26:34.865205 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:34 crc kubenswrapper[4725]: I1202 13:26:34.929880 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-62xsd"] Dec 02 13:26:34 crc kubenswrapper[4725]: I1202 13:26:34.930315 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerName="dnsmasq-dns" containerID="cri-o://e930f4e708ae92c1a8e99fe114e6486ddfe3d54271084236ff0a69c1c07e4251" gracePeriod=10 Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.225334 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55478c4467-chnnr"] Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.229374 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.262801 4725 generic.go:334] "Generic (PLEG): container finished" podID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerID="e930f4e708ae92c1a8e99fe114e6486ddfe3d54271084236ff0a69c1c07e4251" exitCode=0 Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.262857 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" event={"ID":"f5bec6b9-6b89-4569-a34c-ec75f0027ef6","Type":"ContainerDied","Data":"e930f4e708ae92c1a8e99fe114e6486ddfe3d54271084236ff0a69c1c07e4251"} Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.263271 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-chnnr"] Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.312707 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.312750 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-config\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.312797 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.312839 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbz7v\" (UniqueName: \"kubernetes.io/projected/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-kube-api-access-vbz7v\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.312984 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-dns-svc\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.313012 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.313055 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415436 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415528 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415555 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-config\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415593 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415623 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbz7v\" (UniqueName: \"kubernetes.io/projected/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-kube-api-access-vbz7v\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415709 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-dns-svc\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.415733 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.416303 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-dns-swift-storage-0\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.416704 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-dns-svc\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.416803 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-ovsdbserver-sb\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.416962 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-ovsdbserver-nb\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.417288 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-openstack-edpm-ipam\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.417345 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-config\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.439727 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbz7v\" (UniqueName: \"kubernetes.io/projected/cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61-kube-api-access-vbz7v\") pod \"dnsmasq-dns-55478c4467-chnnr\" (UID: \"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61\") " pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.509210 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.564007 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.618483 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-svc\") pod \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.618564 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-config\") pod \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.618597 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-swift-storage-0\") pod \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.618778 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5lbz\" (UniqueName: \"kubernetes.io/projected/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-kube-api-access-p5lbz\") pod \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.619202 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-nb\") pod \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.619267 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-sb\") pod \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\" (UID: \"f5bec6b9-6b89-4569-a34c-ec75f0027ef6\") " Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.622645 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-kube-api-access-p5lbz" (OuterVolumeSpecName: "kube-api-access-p5lbz") pod "f5bec6b9-6b89-4569-a34c-ec75f0027ef6" (UID: "f5bec6b9-6b89-4569-a34c-ec75f0027ef6"). InnerVolumeSpecName "kube-api-access-p5lbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.692415 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f5bec6b9-6b89-4569-a34c-ec75f0027ef6" (UID: "f5bec6b9-6b89-4569-a34c-ec75f0027ef6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.707993 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-config" (OuterVolumeSpecName: "config") pod "f5bec6b9-6b89-4569-a34c-ec75f0027ef6" (UID: "f5bec6b9-6b89-4569-a34c-ec75f0027ef6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.710967 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5bec6b9-6b89-4569-a34c-ec75f0027ef6" (UID: "f5bec6b9-6b89-4569-a34c-ec75f0027ef6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.714924 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f5bec6b9-6b89-4569-a34c-ec75f0027ef6" (UID: "f5bec6b9-6b89-4569-a34c-ec75f0027ef6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.716086 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5bec6b9-6b89-4569-a34c-ec75f0027ef6" (UID: "f5bec6b9-6b89-4569-a34c-ec75f0027ef6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.723171 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5lbz\" (UniqueName: \"kubernetes.io/projected/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-kube-api-access-p5lbz\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.723205 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.723216 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.723235 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.723249 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:35 crc kubenswrapper[4725]: I1202 13:26:35.723261 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f5bec6b9-6b89-4569-a34c-ec75f0027ef6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.012244 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55478c4467-chnnr"] Dec 02 13:26:36 crc kubenswrapper[4725]: W1202 13:26:36.015003 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf142ef8_d0b8_440a_ae2c_7f6f0abb7a61.slice/crio-5ee2df7975889def8063b4d2fc6620d625b6f4b804eba7ecfa94f0546c3dae87 WatchSource:0}: Error finding container 5ee2df7975889def8063b4d2fc6620d625b6f4b804eba7ecfa94f0546c3dae87: Status 404 returned error can't find the container with id 5ee2df7975889def8063b4d2fc6620d625b6f4b804eba7ecfa94f0546c3dae87 Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.277371 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-chnnr" event={"ID":"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61","Type":"ContainerStarted","Data":"37c486951c57a6b24d3660b6db6094f07350ba70ef8938a4c6197f02fc3db683"} Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.277410 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-chnnr" event={"ID":"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61","Type":"ContainerStarted","Data":"5ee2df7975889def8063b4d2fc6620d625b6f4b804eba7ecfa94f0546c3dae87"} Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.280692 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" event={"ID":"f5bec6b9-6b89-4569-a34c-ec75f0027ef6","Type":"ContainerDied","Data":"9dcc42c22d1e51248c995118b51088bc9f20705d8108b8102055f5c1c1e5bad4"} Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.280747 4725 scope.go:117] "RemoveContainer" containerID="e930f4e708ae92c1a8e99fe114e6486ddfe3d54271084236ff0a69c1c07e4251" Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.280907 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-62xsd" Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.474572 4725 scope.go:117] "RemoveContainer" containerID="2db1493e5faad9e9a103bfee3045f3c3a26f0c1d928a7b6ebe08b895ca9b1dc2" Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.479368 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-62xsd"] Dec 02 13:26:36 crc kubenswrapper[4725]: I1202 13:26:36.489001 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-62xsd"] Dec 02 13:26:37 crc kubenswrapper[4725]: I1202 13:26:37.289413 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" path="/var/lib/kubelet/pods/f5bec6b9-6b89-4569-a34c-ec75f0027ef6/volumes" Dec 02 13:26:37 crc kubenswrapper[4725]: I1202 13:26:37.294660 4725 generic.go:334] "Generic (PLEG): container finished" podID="cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61" containerID="37c486951c57a6b24d3660b6db6094f07350ba70ef8938a4c6197f02fc3db683" exitCode=0 Dec 02 13:26:37 crc kubenswrapper[4725]: I1202 13:26:37.294761 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-chnnr" event={"ID":"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61","Type":"ContainerDied","Data":"37c486951c57a6b24d3660b6db6094f07350ba70ef8938a4c6197f02fc3db683"} Dec 02 13:26:37 crc kubenswrapper[4725]: I1202 13:26:37.294822 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55478c4467-chnnr" event={"ID":"cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61","Type":"ContainerStarted","Data":"168295abc7222a38e2bbfa58b8f4a4b500521dbc1d0f84e0b2063871f32a75bc"} Dec 02 13:26:37 crc kubenswrapper[4725]: I1202 13:26:37.295106 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:37 crc kubenswrapper[4725]: I1202 13:26:37.314212 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55478c4467-chnnr" podStartSLOduration=2.314195709 podStartE2EDuration="2.314195709s" podCreationTimestamp="2025-12-02 13:26:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:26:37.313655255 +0000 UTC m=+1328.270296950" watchObservedRunningTime="2025-12-02 13:26:37.314195709 +0000 UTC m=+1328.270837404" Dec 02 13:26:45 crc kubenswrapper[4725]: I1202 13:26:45.565634 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55478c4467-chnnr" Dec 02 13:26:45 crc kubenswrapper[4725]: I1202 13:26:45.631660 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h57r8"] Dec 02 13:26:45 crc kubenswrapper[4725]: I1202 13:26:45.632155 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerName="dnsmasq-dns" containerID="cri-o://a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c" gracePeriod=10 Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.194279 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232171 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-swift-storage-0\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232245 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-config\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232394 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljr5v\" (UniqueName: \"kubernetes.io/projected/ef8dac21-8448-49a5-bc91-506b522da6d4-kube-api-access-ljr5v\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232443 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-sb\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232480 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-svc\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232559 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-nb\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.232578 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-openstack-edpm-ipam\") pod \"ef8dac21-8448-49a5-bc91-506b522da6d4\" (UID: \"ef8dac21-8448-49a5-bc91-506b522da6d4\") " Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.240686 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef8dac21-8448-49a5-bc91-506b522da6d4-kube-api-access-ljr5v" (OuterVolumeSpecName: "kube-api-access-ljr5v") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "kube-api-access-ljr5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.304252 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.308217 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.313971 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-config" (OuterVolumeSpecName: "config") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.324683 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.324711 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.335252 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljr5v\" (UniqueName: \"kubernetes.io/projected/ef8dac21-8448-49a5-bc91-506b522da6d4-kube-api-access-ljr5v\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.335296 4725 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.335307 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.335317 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.335327 4725 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.335336 4725 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-config\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.341211 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef8dac21-8448-49a5-bc91-506b522da6d4" (UID: "ef8dac21-8448-49a5-bc91-506b522da6d4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.383877 4725 generic.go:334] "Generic (PLEG): container finished" podID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerID="a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c" exitCode=0 Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.383922 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" event={"ID":"ef8dac21-8448-49a5-bc91-506b522da6d4","Type":"ContainerDied","Data":"a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c"} Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.383954 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.383976 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-h57r8" event={"ID":"ef8dac21-8448-49a5-bc91-506b522da6d4","Type":"ContainerDied","Data":"5d486096983b3a996aa0904fc79e6c3717ce80efff50dd7419713f7d696c1502"} Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.383995 4725 scope.go:117] "RemoveContainer" containerID="a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.412730 4725 scope.go:117] "RemoveContainer" containerID="126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.420141 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h57r8"] Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.431302 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-h57r8"] Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.437258 4725 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef8dac21-8448-49a5-bc91-506b522da6d4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.448851 4725 scope.go:117] "RemoveContainer" containerID="a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c" Dec 02 13:26:46 crc kubenswrapper[4725]: E1202 13:26:46.451664 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c\": container with ID starting with a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c not found: ID does not exist" containerID="a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.451711 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c"} err="failed to get container status \"a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c\": rpc error: code = NotFound desc = could not find container \"a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c\": container with ID starting with a53f51882fe9e0bbc3c75b7211c14fbfb251d613284007bcafacc6c8a2992f9c not found: ID does not exist" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.451743 4725 scope.go:117] "RemoveContainer" containerID="126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092" Dec 02 13:26:46 crc kubenswrapper[4725]: E1202 13:26:46.454176 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092\": container with ID starting with 126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092 not found: ID does not exist" containerID="126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092" Dec 02 13:26:46 crc kubenswrapper[4725]: I1202 13:26:46.454235 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092"} err="failed to get container status \"126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092\": rpc error: code = NotFound desc = could not find container \"126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092\": container with ID starting with 126ecf75e58b80f577d7fec83936c26bd2f254d3792f3846800e3a309f600092 not found: ID does not exist" Dec 02 13:26:47 crc kubenswrapper[4725]: I1202 13:26:47.280727 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" path="/var/lib/kubelet/pods/ef8dac21-8448-49a5-bc91-506b522da6d4/volumes" Dec 02 13:26:54 crc kubenswrapper[4725]: I1202 13:26:54.285033 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:26:54 crc kubenswrapper[4725]: I1202 13:26:54.285665 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:26:54 crc kubenswrapper[4725]: I1202 13:26:54.285716 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:26:54 crc kubenswrapper[4725]: I1202 13:26:54.286546 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05f4c4672c17d44bd1b7bcedcce35ce67bcd6b4893164e4887e1ed3ea23f75fe"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:26:54 crc kubenswrapper[4725]: I1202 13:26:54.286690 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://05f4c4672c17d44bd1b7bcedcce35ce67bcd6b4893164e4887e1ed3ea23f75fe" gracePeriod=600 Dec 02 13:26:55 crc kubenswrapper[4725]: I1202 13:26:55.326904 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="05f4c4672c17d44bd1b7bcedcce35ce67bcd6b4893164e4887e1ed3ea23f75fe" exitCode=0 Dec 02 13:26:55 crc kubenswrapper[4725]: I1202 13:26:55.327008 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"05f4c4672c17d44bd1b7bcedcce35ce67bcd6b4893164e4887e1ed3ea23f75fe"} Dec 02 13:26:55 crc kubenswrapper[4725]: I1202 13:26:55.327262 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799"} Dec 02 13:26:55 crc kubenswrapper[4725]: I1202 13:26:55.327291 4725 scope.go:117] "RemoveContainer" containerID="1cf7371c50f3e4f7f952b1130b75faf18d2e6023f1c852348322f64d11754fac" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.358179 4725 generic.go:334] "Generic (PLEG): container finished" podID="d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7" containerID="02fe3eb0b73623579b7dc3d091d78fdcb9cf2f7ae0dc68a92b1bc6a7cf0bd957" exitCode=0 Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.358265 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7","Type":"ContainerDied","Data":"02fe3eb0b73623579b7dc3d091d78fdcb9cf2f7ae0dc68a92b1bc6a7cf0bd957"} Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.361131 4725 generic.go:334] "Generic (PLEG): container finished" podID="1774e612-2433-4536-a07c-ef5ac77f2036" containerID="495a729d20fb9b3efe34bd15ac2696d0283994fd8f04dbba1802bd4c797cd743" exitCode=0 Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.361155 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1774e612-2433-4536-a07c-ef5ac77f2036","Type":"ContainerDied","Data":"495a729d20fb9b3efe34bd15ac2696d0283994fd8f04dbba1802bd4c797cd743"} Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.760955 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h"] Dec 02 13:26:58 crc kubenswrapper[4725]: E1202 13:26:58.761842 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerName="dnsmasq-dns" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.761865 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerName="dnsmasq-dns" Dec 02 13:26:58 crc kubenswrapper[4725]: E1202 13:26:58.761893 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerName="init" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.761902 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerName="init" Dec 02 13:26:58 crc kubenswrapper[4725]: E1202 13:26:58.761926 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerName="init" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.761934 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerName="init" Dec 02 13:26:58 crc kubenswrapper[4725]: E1202 13:26:58.761947 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerName="dnsmasq-dns" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.761954 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerName="dnsmasq-dns" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.762195 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8dac21-8448-49a5-bc91-506b522da6d4" containerName="dnsmasq-dns" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.762220 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5bec6b9-6b89-4569-a34c-ec75f0027ef6" containerName="dnsmasq-dns" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.763033 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.767836 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.767895 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.767953 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf6ld\" (UniqueName: \"kubernetes.io/projected/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-kube-api-access-pf6ld\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.767992 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.780777 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.780940 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.781184 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.781287 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.790986 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h"] Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.870380 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.870445 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.870534 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf6ld\" (UniqueName: \"kubernetes.io/projected/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-kube-api-access-pf6ld\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.870591 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.875840 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.876401 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.877606 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:58 crc kubenswrapper[4725]: I1202 13:26:58.890816 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf6ld\" (UniqueName: \"kubernetes.io/projected/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-kube-api-access-pf6ld\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.090899 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.378354 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7","Type":"ContainerStarted","Data":"597ae8ed15230e03a7c480bb36bede71510c6ed3b5f87cea7bde882673125ee7"} Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.379980 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.384925 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1774e612-2433-4536-a07c-ef5ac77f2036","Type":"ContainerStarted","Data":"785e9ed1adba47957d488b152749d0a74539909bedadc336d50f2496c1820479"} Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.385720 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.460025 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.459982415 podStartE2EDuration="36.459982415s" podCreationTimestamp="2025-12-02 13:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:26:59.449957803 +0000 UTC m=+1350.406599508" watchObservedRunningTime="2025-12-02 13:26:59.459982415 +0000 UTC m=+1350.416624110" Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.468875 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.468851478 podStartE2EDuration="37.468851478s" podCreationTimestamp="2025-12-02 13:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 13:26:59.421502798 +0000 UTC m=+1350.378144513" watchObservedRunningTime="2025-12-02 13:26:59.468851478 +0000 UTC m=+1350.425493173" Dec 02 13:26:59 crc kubenswrapper[4725]: W1202 13:26:59.819554 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8994bc56_f6d8_49d0_8236_eab3fe2d35e7.slice/crio-7ab20ef1987b969b4be926094ca091063323456da8f9eec2d473f27800351c34 WatchSource:0}: Error finding container 7ab20ef1987b969b4be926094ca091063323456da8f9eec2d473f27800351c34: Status 404 returned error can't find the container with id 7ab20ef1987b969b4be926094ca091063323456da8f9eec2d473f27800351c34 Dec 02 13:26:59 crc kubenswrapper[4725]: I1202 13:26:59.820931 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h"] Dec 02 13:27:00 crc kubenswrapper[4725]: I1202 13:27:00.396300 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" event={"ID":"8994bc56-f6d8-49d0-8236-eab3fe2d35e7","Type":"ContainerStarted","Data":"7ab20ef1987b969b4be926094ca091063323456da8f9eec2d473f27800351c34"} Dec 02 13:27:11 crc kubenswrapper[4725]: I1202 13:27:11.528404 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" event={"ID":"8994bc56-f6d8-49d0-8236-eab3fe2d35e7","Type":"ContainerStarted","Data":"4bca3d2d998d1b3a540058986063af9178be114257d9a6b8f59ae25357c20675"} Dec 02 13:27:11 crc kubenswrapper[4725]: I1202 13:27:11.547986 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" podStartSLOduration=2.2212223030000002 podStartE2EDuration="13.547960078s" podCreationTimestamp="2025-12-02 13:26:58 +0000 UTC" firstStartedPulling="2025-12-02 13:26:59.822734975 +0000 UTC m=+1350.779376670" lastFinishedPulling="2025-12-02 13:27:11.14947274 +0000 UTC m=+1362.106114445" observedRunningTime="2025-12-02 13:27:11.543407884 +0000 UTC m=+1362.500049579" watchObservedRunningTime="2025-12-02 13:27:11.547960078 +0000 UTC m=+1362.504601773" Dec 02 13:27:12 crc kubenswrapper[4725]: I1202 13:27:12.611620 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 02 13:27:13 crc kubenswrapper[4725]: I1202 13:27:13.647684 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 02 13:27:22 crc kubenswrapper[4725]: I1202 13:27:22.631942 4725 generic.go:334] "Generic (PLEG): container finished" podID="8994bc56-f6d8-49d0-8236-eab3fe2d35e7" containerID="4bca3d2d998d1b3a540058986063af9178be114257d9a6b8f59ae25357c20675" exitCode=0 Dec 02 13:27:22 crc kubenswrapper[4725]: I1202 13:27:22.631964 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" event={"ID":"8994bc56-f6d8-49d0-8236-eab3fe2d35e7","Type":"ContainerDied","Data":"4bca3d2d998d1b3a540058986063af9178be114257d9a6b8f59ae25357c20675"} Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.029027 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.208738 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-inventory\") pod \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.208851 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-repo-setup-combined-ca-bundle\") pod \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.208911 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-ssh-key\") pod \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.208989 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf6ld\" (UniqueName: \"kubernetes.io/projected/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-kube-api-access-pf6ld\") pod \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\" (UID: \"8994bc56-f6d8-49d0-8236-eab3fe2d35e7\") " Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.214796 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8994bc56-f6d8-49d0-8236-eab3fe2d35e7" (UID: "8994bc56-f6d8-49d0-8236-eab3fe2d35e7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.215402 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-kube-api-access-pf6ld" (OuterVolumeSpecName: "kube-api-access-pf6ld") pod "8994bc56-f6d8-49d0-8236-eab3fe2d35e7" (UID: "8994bc56-f6d8-49d0-8236-eab3fe2d35e7"). InnerVolumeSpecName "kube-api-access-pf6ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.237317 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8994bc56-f6d8-49d0-8236-eab3fe2d35e7" (UID: "8994bc56-f6d8-49d0-8236-eab3fe2d35e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.249381 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-inventory" (OuterVolumeSpecName: "inventory") pod "8994bc56-f6d8-49d0-8236-eab3fe2d35e7" (UID: "8994bc56-f6d8-49d0-8236-eab3fe2d35e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.311778 4725 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.311818 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.311830 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf6ld\" (UniqueName: \"kubernetes.io/projected/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-kube-api-access-pf6ld\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.311840 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8994bc56-f6d8-49d0-8236-eab3fe2d35e7-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.654511 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" event={"ID":"8994bc56-f6d8-49d0-8236-eab3fe2d35e7","Type":"ContainerDied","Data":"7ab20ef1987b969b4be926094ca091063323456da8f9eec2d473f27800351c34"} Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.654798 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ab20ef1987b969b4be926094ca091063323456da8f9eec2d473f27800351c34" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.654599 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.730407 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg"] Dec 02 13:27:24 crc kubenswrapper[4725]: E1202 13:27:24.730882 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8994bc56-f6d8-49d0-8236-eab3fe2d35e7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.730901 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="8994bc56-f6d8-49d0-8236-eab3fe2d35e7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.731114 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="8994bc56-f6d8-49d0-8236-eab3fe2d35e7" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.731772 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.734206 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.735212 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.735307 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.741095 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.742649 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg"] Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.922987 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgsb8\" (UniqueName: \"kubernetes.io/projected/efdcb239-8753-47be-86b7-267371f72e53-kube-api-access-qgsb8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.923057 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:24 crc kubenswrapper[4725]: I1202 13:27:24.923192 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.025255 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgsb8\" (UniqueName: \"kubernetes.io/projected/efdcb239-8753-47be-86b7-267371f72e53-kube-api-access-qgsb8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.025328 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.025516 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.034741 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.035102 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.044958 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgsb8\" (UniqueName: \"kubernetes.io/projected/efdcb239-8753-47be-86b7-267371f72e53-kube-api-access-qgsb8\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mjvdg\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.048071 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.609932 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg"] Dec 02 13:27:25 crc kubenswrapper[4725]: W1202 13:27:25.614413 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefdcb239_8753_47be_86b7_267371f72e53.slice/crio-bfaf581cd143eff968989a9d823424683894e3cf5cf2a23334b9870798b6753e WatchSource:0}: Error finding container bfaf581cd143eff968989a9d823424683894e3cf5cf2a23334b9870798b6753e: Status 404 returned error can't find the container with id bfaf581cd143eff968989a9d823424683894e3cf5cf2a23334b9870798b6753e Dec 02 13:27:25 crc kubenswrapper[4725]: I1202 13:27:25.665451 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" event={"ID":"efdcb239-8753-47be-86b7-267371f72e53","Type":"ContainerStarted","Data":"bfaf581cd143eff968989a9d823424683894e3cf5cf2a23334b9870798b6753e"} Dec 02 13:27:26 crc kubenswrapper[4725]: I1202 13:27:26.676002 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" event={"ID":"efdcb239-8753-47be-86b7-267371f72e53","Type":"ContainerStarted","Data":"f3940d5811da415a1bc2157ae65f99cc4fca4f7a427ff2c0988583fc6b83bee7"} Dec 02 13:27:26 crc kubenswrapper[4725]: I1202 13:27:26.691534 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" podStartSLOduration=2.185426129 podStartE2EDuration="2.691509145s" podCreationTimestamp="2025-12-02 13:27:24 +0000 UTC" firstStartedPulling="2025-12-02 13:27:25.616905294 +0000 UTC m=+1376.573546979" lastFinishedPulling="2025-12-02 13:27:26.12298831 +0000 UTC m=+1377.079629995" observedRunningTime="2025-12-02 13:27:26.690253634 +0000 UTC m=+1377.646895329" watchObservedRunningTime="2025-12-02 13:27:26.691509145 +0000 UTC m=+1377.648150840" Dec 02 13:27:29 crc kubenswrapper[4725]: I1202 13:27:29.711158 4725 generic.go:334] "Generic (PLEG): container finished" podID="efdcb239-8753-47be-86b7-267371f72e53" containerID="f3940d5811da415a1bc2157ae65f99cc4fca4f7a427ff2c0988583fc6b83bee7" exitCode=0 Dec 02 13:27:29 crc kubenswrapper[4725]: I1202 13:27:29.711241 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" event={"ID":"efdcb239-8753-47be-86b7-267371f72e53","Type":"ContainerDied","Data":"f3940d5811da415a1bc2157ae65f99cc4fca4f7a427ff2c0988583fc6b83bee7"} Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.287779 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.445824 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-ssh-key\") pod \"efdcb239-8753-47be-86b7-267371f72e53\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.446304 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgsb8\" (UniqueName: \"kubernetes.io/projected/efdcb239-8753-47be-86b7-267371f72e53-kube-api-access-qgsb8\") pod \"efdcb239-8753-47be-86b7-267371f72e53\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.446381 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-inventory\") pod \"efdcb239-8753-47be-86b7-267371f72e53\" (UID: \"efdcb239-8753-47be-86b7-267371f72e53\") " Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.451492 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdcb239-8753-47be-86b7-267371f72e53-kube-api-access-qgsb8" (OuterVolumeSpecName: "kube-api-access-qgsb8") pod "efdcb239-8753-47be-86b7-267371f72e53" (UID: "efdcb239-8753-47be-86b7-267371f72e53"). InnerVolumeSpecName "kube-api-access-qgsb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.479177 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-inventory" (OuterVolumeSpecName: "inventory") pod "efdcb239-8753-47be-86b7-267371f72e53" (UID: "efdcb239-8753-47be-86b7-267371f72e53"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.479938 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "efdcb239-8753-47be-86b7-267371f72e53" (UID: "efdcb239-8753-47be-86b7-267371f72e53"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.549231 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.549276 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efdcb239-8753-47be-86b7-267371f72e53-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.549290 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgsb8\" (UniqueName: \"kubernetes.io/projected/efdcb239-8753-47be-86b7-267371f72e53-kube-api-access-qgsb8\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.728576 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" event={"ID":"efdcb239-8753-47be-86b7-267371f72e53","Type":"ContainerDied","Data":"bfaf581cd143eff968989a9d823424683894e3cf5cf2a23334b9870798b6753e"} Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.728981 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfaf581cd143eff968989a9d823424683894e3cf5cf2a23334b9870798b6753e" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.728664 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mjvdg" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.805867 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr"] Dec 02 13:27:31 crc kubenswrapper[4725]: E1202 13:27:31.806332 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdcb239-8753-47be-86b7-267371f72e53" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.806349 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdcb239-8753-47be-86b7-267371f72e53" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.806555 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="efdcb239-8753-47be-86b7-267371f72e53" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.807268 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.809970 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.810504 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.810732 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.812366 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.824689 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr"] Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.957490 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.957527 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.958123 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcdll\" (UniqueName: \"kubernetes.io/projected/3ab59324-3446-4602-af02-e9094f220f34-kube-api-access-xcdll\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:31 crc kubenswrapper[4725]: I1202 13:27:31.958245 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.060651 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcdll\" (UniqueName: \"kubernetes.io/projected/3ab59324-3446-4602-af02-e9094f220f34-kube-api-access-xcdll\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.060800 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.060969 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.061006 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.066088 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.066301 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.067068 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.079602 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcdll\" (UniqueName: \"kubernetes.io/projected/3ab59324-3446-4602-af02-e9094f220f34-kube-api-access-xcdll\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.134570 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.624630 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr"] Dec 02 13:27:32 crc kubenswrapper[4725]: W1202 13:27:32.627963 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ab59324_3446_4602_af02_e9094f220f34.slice/crio-8ba76f9ebaa74940c0ac64c6e036470e01ba1b3d4b855203eecf60dc8813c74b WatchSource:0}: Error finding container 8ba76f9ebaa74940c0ac64c6e036470e01ba1b3d4b855203eecf60dc8813c74b: Status 404 returned error can't find the container with id 8ba76f9ebaa74940c0ac64c6e036470e01ba1b3d4b855203eecf60dc8813c74b Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.739304 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" event={"ID":"3ab59324-3446-4602-af02-e9094f220f34","Type":"ContainerStarted","Data":"8ba76f9ebaa74940c0ac64c6e036470e01ba1b3d4b855203eecf60dc8813c74b"} Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.794948 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8pnw"] Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.796802 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.812122 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8pnw"] Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.878979 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-catalog-content\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.879225 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7gsf\" (UniqueName: \"kubernetes.io/projected/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-kube-api-access-r7gsf\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.879262 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-utilities\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.981505 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7gsf\" (UniqueName: \"kubernetes.io/projected/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-kube-api-access-r7gsf\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.981569 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-utilities\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.981598 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-catalog-content\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.982274 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-utilities\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:32 crc kubenswrapper[4725]: I1202 13:27:32.982317 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-catalog-content\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:33 crc kubenswrapper[4725]: I1202 13:27:33.002485 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7gsf\" (UniqueName: \"kubernetes.io/projected/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-kube-api-access-r7gsf\") pod \"redhat-marketplace-b8pnw\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:33 crc kubenswrapper[4725]: I1202 13:27:33.140441 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:33 crc kubenswrapper[4725]: I1202 13:27:33.607833 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8pnw"] Dec 02 13:27:33 crc kubenswrapper[4725]: I1202 13:27:33.749534 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" event={"ID":"3ab59324-3446-4602-af02-e9094f220f34","Type":"ContainerStarted","Data":"28b80448f9548a67732564e76303892c724cab4fe81ba0a4325582f91da04604"} Dec 02 13:27:33 crc kubenswrapper[4725]: I1202 13:27:33.754443 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerStarted","Data":"f4e30f910c1dd58e68f03404ea1233b06b74cc998ee555f6eee4298eb6be6391"} Dec 02 13:27:33 crc kubenswrapper[4725]: I1202 13:27:33.775016 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" podStartSLOduration=2.214767077 podStartE2EDuration="2.774994965s" podCreationTimestamp="2025-12-02 13:27:31 +0000 UTC" firstStartedPulling="2025-12-02 13:27:32.630409986 +0000 UTC m=+1383.587051681" lastFinishedPulling="2025-12-02 13:27:33.190637874 +0000 UTC m=+1384.147279569" observedRunningTime="2025-12-02 13:27:33.763282442 +0000 UTC m=+1384.719924177" watchObservedRunningTime="2025-12-02 13:27:33.774994965 +0000 UTC m=+1384.731636670" Dec 02 13:27:34 crc kubenswrapper[4725]: I1202 13:27:34.765201 4725 generic.go:334] "Generic (PLEG): container finished" podID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerID="2f46d8863cae3b7595faf9b2e979579633b73b63008a54b957ca0ce9dc6bee18" exitCode=0 Dec 02 13:27:34 crc kubenswrapper[4725]: I1202 13:27:34.765259 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerDied","Data":"2f46d8863cae3b7595faf9b2e979579633b73b63008a54b957ca0ce9dc6bee18"} Dec 02 13:27:35 crc kubenswrapper[4725]: I1202 13:27:35.775917 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerStarted","Data":"40b77c20241ca086c6d9446589a4b907e013982c8ff95fd853ad2cec508cfd4c"} Dec 02 13:27:36 crc kubenswrapper[4725]: I1202 13:27:36.785417 4725 generic.go:334] "Generic (PLEG): container finished" podID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerID="40b77c20241ca086c6d9446589a4b907e013982c8ff95fd853ad2cec508cfd4c" exitCode=0 Dec 02 13:27:36 crc kubenswrapper[4725]: I1202 13:27:36.785510 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerDied","Data":"40b77c20241ca086c6d9446589a4b907e013982c8ff95fd853ad2cec508cfd4c"} Dec 02 13:27:37 crc kubenswrapper[4725]: I1202 13:27:37.798473 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerStarted","Data":"bf9a62d00dd6f0ceb4906e0f581db0961eb660e16074e763ad0730925d8188ee"} Dec 02 13:27:37 crc kubenswrapper[4725]: I1202 13:27:37.819769 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8pnw" podStartSLOduration=3.299824355 podStartE2EDuration="5.819749852s" podCreationTimestamp="2025-12-02 13:27:32 +0000 UTC" firstStartedPulling="2025-12-02 13:27:34.767622958 +0000 UTC m=+1385.724264663" lastFinishedPulling="2025-12-02 13:27:37.287548465 +0000 UTC m=+1388.244190160" observedRunningTime="2025-12-02 13:27:37.818365958 +0000 UTC m=+1388.775007653" watchObservedRunningTime="2025-12-02 13:27:37.819749852 +0000 UTC m=+1388.776391547" Dec 02 13:27:43 crc kubenswrapper[4725]: I1202 13:27:43.142069 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:43 crc kubenswrapper[4725]: I1202 13:27:43.142652 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:43 crc kubenswrapper[4725]: I1202 13:27:43.207713 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:43 crc kubenswrapper[4725]: I1202 13:27:43.907057 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:43 crc kubenswrapper[4725]: I1202 13:27:43.963656 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8pnw"] Dec 02 13:27:45 crc kubenswrapper[4725]: I1202 13:27:45.871124 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b8pnw" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="registry-server" containerID="cri-o://bf9a62d00dd6f0ceb4906e0f581db0961eb660e16074e763ad0730925d8188ee" gracePeriod=2 Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.882408 4725 generic.go:334] "Generic (PLEG): container finished" podID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerID="bf9a62d00dd6f0ceb4906e0f581db0961eb660e16074e763ad0730925d8188ee" exitCode=0 Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.882496 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerDied","Data":"bf9a62d00dd6f0ceb4906e0f581db0961eb660e16074e763ad0730925d8188ee"} Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.883751 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8pnw" event={"ID":"9b49cea7-1826-4fc5-b89e-12d79ff82f1e","Type":"ContainerDied","Data":"f4e30f910c1dd58e68f03404ea1233b06b74cc998ee555f6eee4298eb6be6391"} Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.883812 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4e30f910c1dd58e68f03404ea1233b06b74cc998ee555f6eee4298eb6be6391" Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.890536 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.952700 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-catalog-content\") pod \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.952897 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-utilities\") pod \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.953000 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7gsf\" (UniqueName: \"kubernetes.io/projected/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-kube-api-access-r7gsf\") pod \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\" (UID: \"9b49cea7-1826-4fc5-b89e-12d79ff82f1e\") " Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.953641 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-utilities" (OuterVolumeSpecName: "utilities") pod "9b49cea7-1826-4fc5-b89e-12d79ff82f1e" (UID: "9b49cea7-1826-4fc5-b89e-12d79ff82f1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.959999 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-kube-api-access-r7gsf" (OuterVolumeSpecName: "kube-api-access-r7gsf") pod "9b49cea7-1826-4fc5-b89e-12d79ff82f1e" (UID: "9b49cea7-1826-4fc5-b89e-12d79ff82f1e"). InnerVolumeSpecName "kube-api-access-r7gsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:27:46 crc kubenswrapper[4725]: I1202 13:27:46.971509 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9b49cea7-1826-4fc5-b89e-12d79ff82f1e" (UID: "9b49cea7-1826-4fc5-b89e-12d79ff82f1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.055324 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7gsf\" (UniqueName: \"kubernetes.io/projected/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-kube-api-access-r7gsf\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.055592 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.055665 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9b49cea7-1826-4fc5-b89e-12d79ff82f1e-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.143234 4725 scope.go:117] "RemoveContainer" containerID="115389f2abd70000c9b731354b036fa08a2b1652d50bb65ca5e5328213ac48ea" Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.892422 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8pnw" Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.919626 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8pnw"] Dec 02 13:27:47 crc kubenswrapper[4725]: I1202 13:27:47.929193 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8pnw"] Dec 02 13:27:49 crc kubenswrapper[4725]: I1202 13:27:49.282144 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" path="/var/lib/kubelet/pods/9b49cea7-1826-4fc5-b89e-12d79ff82f1e/volumes" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.249882 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4qzhx"] Dec 02 13:27:59 crc kubenswrapper[4725]: E1202 13:27:59.251145 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="registry-server" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.251166 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="registry-server" Dec 02 13:27:59 crc kubenswrapper[4725]: E1202 13:27:59.251203 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="extract-utilities" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.251213 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="extract-utilities" Dec 02 13:27:59 crc kubenswrapper[4725]: E1202 13:27:59.251230 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="extract-content" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.251239 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="extract-content" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.251546 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b49cea7-1826-4fc5-b89e-12d79ff82f1e" containerName="registry-server" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.253449 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.294845 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4qzhx"] Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.413775 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-catalog-content\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.416166 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl5hm\" (UniqueName: \"kubernetes.io/projected/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-kube-api-access-fl5hm\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.416440 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-utilities\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.519140 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl5hm\" (UniqueName: \"kubernetes.io/projected/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-kube-api-access-fl5hm\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.519495 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-utilities\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.519570 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-catalog-content\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.520112 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-utilities\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.520282 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-catalog-content\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.544057 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl5hm\" (UniqueName: \"kubernetes.io/projected/ea32e5f6-b2b8-49a4-b7a0-2670dd62561b-kube-api-access-fl5hm\") pod \"redhat-operators-4qzhx\" (UID: \"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b\") " pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:27:59 crc kubenswrapper[4725]: I1202 13:27:59.586598 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:28:00 crc kubenswrapper[4725]: I1202 13:28:00.043113 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4qzhx"] Dec 02 13:28:00 crc kubenswrapper[4725]: W1202 13:28:00.048826 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea32e5f6_b2b8_49a4_b7a0_2670dd62561b.slice/crio-3de64f5aa00898fd41d3ddd773f649b4dd9843a259dd548ee513ff52b806e5aa WatchSource:0}: Error finding container 3de64f5aa00898fd41d3ddd773f649b4dd9843a259dd548ee513ff52b806e5aa: Status 404 returned error can't find the container with id 3de64f5aa00898fd41d3ddd773f649b4dd9843a259dd548ee513ff52b806e5aa Dec 02 13:28:01 crc kubenswrapper[4725]: I1202 13:28:01.008677 4725 generic.go:334] "Generic (PLEG): container finished" podID="ea32e5f6-b2b8-49a4-b7a0-2670dd62561b" containerID="28d8f68b7e217f614578bebadfa0c72fee7a5c2f21e067e3d836e917dcefe7d9" exitCode=0 Dec 02 13:28:01 crc kubenswrapper[4725]: I1202 13:28:01.008924 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qzhx" event={"ID":"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b","Type":"ContainerDied","Data":"28d8f68b7e217f614578bebadfa0c72fee7a5c2f21e067e3d836e917dcefe7d9"} Dec 02 13:28:01 crc kubenswrapper[4725]: I1202 13:28:01.008952 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qzhx" event={"ID":"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b","Type":"ContainerStarted","Data":"3de64f5aa00898fd41d3ddd773f649b4dd9843a259dd548ee513ff52b806e5aa"} Dec 02 13:28:11 crc kubenswrapper[4725]: I1202 13:28:11.110149 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qzhx" event={"ID":"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b","Type":"ContainerStarted","Data":"f634e1a603028c1272891f7ce8b84dbd01235381e3861b64b4941a92d57ad2cc"} Dec 02 13:28:13 crc kubenswrapper[4725]: I1202 13:28:13.239376 4725 generic.go:334] "Generic (PLEG): container finished" podID="ea32e5f6-b2b8-49a4-b7a0-2670dd62561b" containerID="f634e1a603028c1272891f7ce8b84dbd01235381e3861b64b4941a92d57ad2cc" exitCode=0 Dec 02 13:28:13 crc kubenswrapper[4725]: I1202 13:28:13.239503 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qzhx" event={"ID":"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b","Type":"ContainerDied","Data":"f634e1a603028c1272891f7ce8b84dbd01235381e3861b64b4941a92d57ad2cc"} Dec 02 13:28:15 crc kubenswrapper[4725]: I1202 13:28:15.262108 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qzhx" event={"ID":"ea32e5f6-b2b8-49a4-b7a0-2670dd62561b","Type":"ContainerStarted","Data":"e9e5a5c63d273534d21abd2653470ab6bfb428f1099a41d85cbb9e320514b4e4"} Dec 02 13:28:15 crc kubenswrapper[4725]: I1202 13:28:15.292491 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4qzhx" podStartSLOduration=2.473646372 podStartE2EDuration="16.292442047s" podCreationTimestamp="2025-12-02 13:27:59 +0000 UTC" firstStartedPulling="2025-12-02 13:28:01.011487032 +0000 UTC m=+1411.968128727" lastFinishedPulling="2025-12-02 13:28:14.830282707 +0000 UTC m=+1425.786924402" observedRunningTime="2025-12-02 13:28:15.285698018 +0000 UTC m=+1426.242339713" watchObservedRunningTime="2025-12-02 13:28:15.292442047 +0000 UTC m=+1426.249083752" Dec 02 13:28:19 crc kubenswrapper[4725]: I1202 13:28:19.587143 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:28:19 crc kubenswrapper[4725]: I1202 13:28:19.587431 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:28:19 crc kubenswrapper[4725]: I1202 13:28:19.633428 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.353307 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4qzhx" Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.417113 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4qzhx"] Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.456568 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8mhmp"] Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.456939 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8mhmp" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="registry-server" containerID="cri-o://c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692" gracePeriod=2 Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.917535 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.992483 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/1ef8608f-fa22-48e6-858b-c23693d8133e-kube-api-access-x5xkd\") pod \"1ef8608f-fa22-48e6-858b-c23693d8133e\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.992600 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-catalog-content\") pod \"1ef8608f-fa22-48e6-858b-c23693d8133e\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.992686 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-utilities\") pod \"1ef8608f-fa22-48e6-858b-c23693d8133e\" (UID: \"1ef8608f-fa22-48e6-858b-c23693d8133e\") " Dec 02 13:28:20 crc kubenswrapper[4725]: I1202 13:28:20.993477 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-utilities" (OuterVolumeSpecName: "utilities") pod "1ef8608f-fa22-48e6-858b-c23693d8133e" (UID: "1ef8608f-fa22-48e6-858b-c23693d8133e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.004766 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef8608f-fa22-48e6-858b-c23693d8133e-kube-api-access-x5xkd" (OuterVolumeSpecName: "kube-api-access-x5xkd") pod "1ef8608f-fa22-48e6-858b-c23693d8133e" (UID: "1ef8608f-fa22-48e6-858b-c23693d8133e"). InnerVolumeSpecName "kube-api-access-x5xkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.095000 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5xkd\" (UniqueName: \"kubernetes.io/projected/1ef8608f-fa22-48e6-858b-c23693d8133e-kube-api-access-x5xkd\") on node \"crc\" DevicePath \"\"" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.095042 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.114204 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ef8608f-fa22-48e6-858b-c23693d8133e" (UID: "1ef8608f-fa22-48e6-858b-c23693d8133e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.197115 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ef8608f-fa22-48e6-858b-c23693d8133e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.315300 4725 generic.go:334] "Generic (PLEG): container finished" podID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerID="c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692" exitCode=0 Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.315420 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8mhmp" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.315428 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8mhmp" event={"ID":"1ef8608f-fa22-48e6-858b-c23693d8133e","Type":"ContainerDied","Data":"c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692"} Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.315491 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8mhmp" event={"ID":"1ef8608f-fa22-48e6-858b-c23693d8133e","Type":"ContainerDied","Data":"107f368fdf8f65459629fd0630dc265a6b348fcfdbb4b5c153e16aab09906c78"} Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.315518 4725 scope.go:117] "RemoveContainer" containerID="c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.346016 4725 scope.go:117] "RemoveContainer" containerID="a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.347279 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8mhmp"] Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.357256 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8mhmp"] Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.372244 4725 scope.go:117] "RemoveContainer" containerID="60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.430692 4725 scope.go:117] "RemoveContainer" containerID="c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692" Dec 02 13:28:21 crc kubenswrapper[4725]: E1202 13:28:21.431451 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692\": container with ID starting with c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692 not found: ID does not exist" containerID="c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.431517 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692"} err="failed to get container status \"c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692\": rpc error: code = NotFound desc = could not find container \"c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692\": container with ID starting with c0f3903d896573ebabad479636e995b020bd6951de018edcecd31fc4420db692 not found: ID does not exist" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.431546 4725 scope.go:117] "RemoveContainer" containerID="a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c" Dec 02 13:28:21 crc kubenswrapper[4725]: E1202 13:28:21.431999 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c\": container with ID starting with a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c not found: ID does not exist" containerID="a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.432029 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c"} err="failed to get container status \"a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c\": rpc error: code = NotFound desc = could not find container \"a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c\": container with ID starting with a468cb9526ec100f923904881bdf65c98e29d48e09bc50900bf0a98d4314045c not found: ID does not exist" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.432050 4725 scope.go:117] "RemoveContainer" containerID="60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b" Dec 02 13:28:21 crc kubenswrapper[4725]: E1202 13:28:21.432296 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b\": container with ID starting with 60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b not found: ID does not exist" containerID="60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b" Dec 02 13:28:21 crc kubenswrapper[4725]: I1202 13:28:21.432324 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b"} err="failed to get container status \"60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b\": rpc error: code = NotFound desc = could not find container \"60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b\": container with ID starting with 60e7979a4bda562f02bbc7d1b8735c320ba2eba3aa7c1d690aa50c71c47cc26b not found: ID does not exist" Dec 02 13:28:23 crc kubenswrapper[4725]: I1202 13:28:23.280263 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" path="/var/lib/kubelet/pods/1ef8608f-fa22-48e6-858b-c23693d8133e/volumes" Dec 02 13:28:47 crc kubenswrapper[4725]: I1202 13:28:47.213951 4725 scope.go:117] "RemoveContainer" containerID="fd8145ab7dcf6d7a1d56c5b97a382302a90ddece38539fefd1b7a4258e63a35a" Dec 02 13:28:47 crc kubenswrapper[4725]: I1202 13:28:47.254130 4725 scope.go:117] "RemoveContainer" containerID="aa9195c8c0f9f6caee04b69871fa844b4d1778256decd817af290cf37baa99ed" Dec 02 13:28:47 crc kubenswrapper[4725]: I1202 13:28:47.285426 4725 scope.go:117] "RemoveContainer" containerID="d5081f4fe2afc34f7dfe0e3e6b66d91b62ae81f8fb97d28f93ff7676fb780338" Dec 02 13:28:47 crc kubenswrapper[4725]: I1202 13:28:47.317798 4725 scope.go:117] "RemoveContainer" containerID="4ab58d5470d7602b49cc92f4e2bcbf04039ed1a49a679485db252829c8ab31cf" Dec 02 13:28:47 crc kubenswrapper[4725]: I1202 13:28:47.350708 4725 scope.go:117] "RemoveContainer" containerID="571e8efba2a1c18cd0ecca1e18b0d5417ac5245d5e2d9b9108517b42b732778a" Dec 02 13:28:54 crc kubenswrapper[4725]: I1202 13:28:54.284819 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:28:54 crc kubenswrapper[4725]: I1202 13:28:54.285388 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:29:24 crc kubenswrapper[4725]: I1202 13:29:24.284501 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:29:24 crc kubenswrapper[4725]: I1202 13:29:24.286645 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.836556 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7zzcl"] Dec 02 13:29:29 crc kubenswrapper[4725]: E1202 13:29:29.837478 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="extract-content" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.837494 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="extract-content" Dec 02 13:29:29 crc kubenswrapper[4725]: E1202 13:29:29.837526 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="registry-server" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.837534 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="registry-server" Dec 02 13:29:29 crc kubenswrapper[4725]: E1202 13:29:29.837554 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="extract-utilities" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.837561 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="extract-utilities" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.837731 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef8608f-fa22-48e6-858b-c23693d8133e" containerName="registry-server" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.839241 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.863392 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwkrb\" (UniqueName: \"kubernetes.io/projected/5305063a-d64e-4cc6-9406-98bb8a99ec38-kube-api-access-gwkrb\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.863705 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-catalog-content\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.863738 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-utilities\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.865586 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zzcl"] Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.965178 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-catalog-content\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.965233 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-utilities\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.965352 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwkrb\" (UniqueName: \"kubernetes.io/projected/5305063a-d64e-4cc6-9406-98bb8a99ec38-kube-api-access-gwkrb\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.965813 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-catalog-content\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.965860 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-utilities\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:29 crc kubenswrapper[4725]: I1202 13:29:29.985975 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwkrb\" (UniqueName: \"kubernetes.io/projected/5305063a-d64e-4cc6-9406-98bb8a99ec38-kube-api-access-gwkrb\") pod \"community-operators-7zzcl\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:30 crc kubenswrapper[4725]: I1202 13:29:30.163139 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:30 crc kubenswrapper[4725]: I1202 13:29:30.734618 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7zzcl"] Dec 02 13:29:31 crc kubenswrapper[4725]: I1202 13:29:31.212765 4725 generic.go:334] "Generic (PLEG): container finished" podID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerID="7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8" exitCode=0 Dec 02 13:29:31 crc kubenswrapper[4725]: I1202 13:29:31.212809 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerDied","Data":"7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8"} Dec 02 13:29:31 crc kubenswrapper[4725]: I1202 13:29:31.212835 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerStarted","Data":"ef1c9516cf6f437d18b6688381c6c50da9a7e229e1acb713f6989afe4898d880"} Dec 02 13:29:33 crc kubenswrapper[4725]: I1202 13:29:33.236833 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerStarted","Data":"ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d"} Dec 02 13:29:34 crc kubenswrapper[4725]: I1202 13:29:34.250795 4725 generic.go:334] "Generic (PLEG): container finished" podID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerID="ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d" exitCode=0 Dec 02 13:29:34 crc kubenswrapper[4725]: I1202 13:29:34.250852 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerDied","Data":"ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d"} Dec 02 13:29:36 crc kubenswrapper[4725]: I1202 13:29:36.283708 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerStarted","Data":"a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b"} Dec 02 13:29:36 crc kubenswrapper[4725]: I1202 13:29:36.334611 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7zzcl" podStartSLOduration=2.983518585 podStartE2EDuration="7.33459017s" podCreationTimestamp="2025-12-02 13:29:29 +0000 UTC" firstStartedPulling="2025-12-02 13:29:31.214746326 +0000 UTC m=+1502.171388021" lastFinishedPulling="2025-12-02 13:29:35.565817911 +0000 UTC m=+1506.522459606" observedRunningTime="2025-12-02 13:29:36.334405786 +0000 UTC m=+1507.291047491" watchObservedRunningTime="2025-12-02 13:29:36.33459017 +0000 UTC m=+1507.291231865" Dec 02 13:29:40 crc kubenswrapper[4725]: I1202 13:29:40.163242 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:40 crc kubenswrapper[4725]: I1202 13:29:40.163783 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:40 crc kubenswrapper[4725]: I1202 13:29:40.226369 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:40 crc kubenswrapper[4725]: I1202 13:29:40.369627 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:40 crc kubenswrapper[4725]: I1202 13:29:40.473629 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zzcl"] Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.334746 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7zzcl" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="registry-server" containerID="cri-o://a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b" gracePeriod=2 Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.799850 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.956315 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-utilities\") pod \"5305063a-d64e-4cc6-9406-98bb8a99ec38\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.956402 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-catalog-content\") pod \"5305063a-d64e-4cc6-9406-98bb8a99ec38\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.956445 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwkrb\" (UniqueName: \"kubernetes.io/projected/5305063a-d64e-4cc6-9406-98bb8a99ec38-kube-api-access-gwkrb\") pod \"5305063a-d64e-4cc6-9406-98bb8a99ec38\" (UID: \"5305063a-d64e-4cc6-9406-98bb8a99ec38\") " Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.957698 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-utilities" (OuterVolumeSpecName: "utilities") pod "5305063a-d64e-4cc6-9406-98bb8a99ec38" (UID: "5305063a-d64e-4cc6-9406-98bb8a99ec38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:29:42 crc kubenswrapper[4725]: I1202 13:29:42.968882 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5305063a-d64e-4cc6-9406-98bb8a99ec38-kube-api-access-gwkrb" (OuterVolumeSpecName: "kube-api-access-gwkrb") pod "5305063a-d64e-4cc6-9406-98bb8a99ec38" (UID: "5305063a-d64e-4cc6-9406-98bb8a99ec38"). InnerVolumeSpecName "kube-api-access-gwkrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.014320 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5305063a-d64e-4cc6-9406-98bb8a99ec38" (UID: "5305063a-d64e-4cc6-9406-98bb8a99ec38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.058997 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.059277 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5305063a-d64e-4cc6-9406-98bb8a99ec38-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.059290 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwkrb\" (UniqueName: \"kubernetes.io/projected/5305063a-d64e-4cc6-9406-98bb8a99ec38-kube-api-access-gwkrb\") on node \"crc\" DevicePath \"\"" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.346067 4725 generic.go:334] "Generic (PLEG): container finished" podID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerID="a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b" exitCode=0 Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.346125 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerDied","Data":"a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b"} Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.346144 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7zzcl" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.346172 4725 scope.go:117] "RemoveContainer" containerID="a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.346158 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7zzcl" event={"ID":"5305063a-d64e-4cc6-9406-98bb8a99ec38","Type":"ContainerDied","Data":"ef1c9516cf6f437d18b6688381c6c50da9a7e229e1acb713f6989afe4898d880"} Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.369357 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7zzcl"] Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.369686 4725 scope.go:117] "RemoveContainer" containerID="ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.379443 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7zzcl"] Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.394906 4725 scope.go:117] "RemoveContainer" containerID="7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.439593 4725 scope.go:117] "RemoveContainer" containerID="a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b" Dec 02 13:29:43 crc kubenswrapper[4725]: E1202 13:29:43.440388 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b\": container with ID starting with a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b not found: ID does not exist" containerID="a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.440449 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b"} err="failed to get container status \"a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b\": rpc error: code = NotFound desc = could not find container \"a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b\": container with ID starting with a181884263abe358533100808307f687381df423e563eac09e48898892b9b77b not found: ID does not exist" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.440522 4725 scope.go:117] "RemoveContainer" containerID="ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d" Dec 02 13:29:43 crc kubenswrapper[4725]: E1202 13:29:43.441139 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d\": container with ID starting with ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d not found: ID does not exist" containerID="ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.441177 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d"} err="failed to get container status \"ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d\": rpc error: code = NotFound desc = could not find container \"ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d\": container with ID starting with ae5b7089f8f8cfb2b2c9592baace4c707d5636404cba4a62287ec9a118e4417d not found: ID does not exist" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.441208 4725 scope.go:117] "RemoveContainer" containerID="7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8" Dec 02 13:29:43 crc kubenswrapper[4725]: E1202 13:29:43.441527 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8\": container with ID starting with 7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8 not found: ID does not exist" containerID="7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8" Dec 02 13:29:43 crc kubenswrapper[4725]: I1202 13:29:43.441557 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8"} err="failed to get container status \"7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8\": rpc error: code = NotFound desc = could not find container \"7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8\": container with ID starting with 7bcd0f4d5b13f93cfed3a2379cabb3164749921553f7c3e91f84e216123fe2c8 not found: ID does not exist" Dec 02 13:29:45 crc kubenswrapper[4725]: I1202 13:29:45.278412 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" path="/var/lib/kubelet/pods/5305063a-d64e-4cc6-9406-98bb8a99ec38/volumes" Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.284561 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.285557 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.285625 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.286805 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.286872 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" gracePeriod=600 Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.453849 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" exitCode=0 Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.453921 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799"} Dec 02 13:29:54 crc kubenswrapper[4725]: I1202 13:29:54.453990 4725 scope.go:117] "RemoveContainer" containerID="05f4c4672c17d44bd1b7bcedcce35ce67bcd6b4893164e4887e1ed3ea23f75fe" Dec 02 13:29:54 crc kubenswrapper[4725]: E1202 13:29:54.464110 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:29:55 crc kubenswrapper[4725]: I1202 13:29:55.471501 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:29:55 crc kubenswrapper[4725]: E1202 13:29:55.472105 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.150913 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz"] Dec 02 13:30:00 crc kubenswrapper[4725]: E1202 13:30:00.152608 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="extract-utilities" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.152627 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="extract-utilities" Dec 02 13:30:00 crc kubenswrapper[4725]: E1202 13:30:00.152646 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="registry-server" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.152653 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="registry-server" Dec 02 13:30:00 crc kubenswrapper[4725]: E1202 13:30:00.152703 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="extract-content" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.152710 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="extract-content" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.152958 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5305063a-d64e-4cc6-9406-98bb8a99ec38" containerName="registry-server" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.153793 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.156493 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.158256 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.172695 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz"] Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.266041 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c51f8b2-f00f-47e4-92ae-000fac82e331-config-volume\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.266406 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c51f8b2-f00f-47e4-92ae-000fac82e331-secret-volume\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.266591 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvlrx\" (UniqueName: \"kubernetes.io/projected/5c51f8b2-f00f-47e4-92ae-000fac82e331-kube-api-access-hvlrx\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.369354 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c51f8b2-f00f-47e4-92ae-000fac82e331-secret-volume\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.369437 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvlrx\" (UniqueName: \"kubernetes.io/projected/5c51f8b2-f00f-47e4-92ae-000fac82e331-kube-api-access-hvlrx\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.369772 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c51f8b2-f00f-47e4-92ae-000fac82e331-config-volume\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.370568 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c51f8b2-f00f-47e4-92ae-000fac82e331-config-volume\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.382296 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c51f8b2-f00f-47e4-92ae-000fac82e331-secret-volume\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.388714 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvlrx\" (UniqueName: \"kubernetes.io/projected/5c51f8b2-f00f-47e4-92ae-000fac82e331-kube-api-access-hvlrx\") pod \"collect-profiles-29411370-2x9tz\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.473281 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:00 crc kubenswrapper[4725]: I1202 13:30:00.943776 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz"] Dec 02 13:30:01 crc kubenswrapper[4725]: I1202 13:30:01.521005 4725 generic.go:334] "Generic (PLEG): container finished" podID="5c51f8b2-f00f-47e4-92ae-000fac82e331" containerID="93ba1471b4e1675c3bf7158abf29ff66ee0f802671eb26a8b191dcfa81f7c4d1" exitCode=0 Dec 02 13:30:01 crc kubenswrapper[4725]: I1202 13:30:01.521085 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" event={"ID":"5c51f8b2-f00f-47e4-92ae-000fac82e331","Type":"ContainerDied","Data":"93ba1471b4e1675c3bf7158abf29ff66ee0f802671eb26a8b191dcfa81f7c4d1"} Dec 02 13:30:01 crc kubenswrapper[4725]: I1202 13:30:01.521614 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" event={"ID":"5c51f8b2-f00f-47e4-92ae-000fac82e331","Type":"ContainerStarted","Data":"38c9d5eae8e821065072584fb5dd9904dac8b19512f6b3ebbbd6f5c2e0978525"} Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.889996 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.927088 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvlrx\" (UniqueName: \"kubernetes.io/projected/5c51f8b2-f00f-47e4-92ae-000fac82e331-kube-api-access-hvlrx\") pod \"5c51f8b2-f00f-47e4-92ae-000fac82e331\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.927263 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c51f8b2-f00f-47e4-92ae-000fac82e331-secret-volume\") pod \"5c51f8b2-f00f-47e4-92ae-000fac82e331\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.927318 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c51f8b2-f00f-47e4-92ae-000fac82e331-config-volume\") pod \"5c51f8b2-f00f-47e4-92ae-000fac82e331\" (UID: \"5c51f8b2-f00f-47e4-92ae-000fac82e331\") " Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.929044 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c51f8b2-f00f-47e4-92ae-000fac82e331-config-volume" (OuterVolumeSpecName: "config-volume") pod "5c51f8b2-f00f-47e4-92ae-000fac82e331" (UID: "5c51f8b2-f00f-47e4-92ae-000fac82e331"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.933988 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c51f8b2-f00f-47e4-92ae-000fac82e331-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.938778 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c51f8b2-f00f-47e4-92ae-000fac82e331-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5c51f8b2-f00f-47e4-92ae-000fac82e331" (UID: "5c51f8b2-f00f-47e4-92ae-000fac82e331"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:30:02 crc kubenswrapper[4725]: I1202 13:30:02.938886 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c51f8b2-f00f-47e4-92ae-000fac82e331-kube-api-access-hvlrx" (OuterVolumeSpecName: "kube-api-access-hvlrx") pod "5c51f8b2-f00f-47e4-92ae-000fac82e331" (UID: "5c51f8b2-f00f-47e4-92ae-000fac82e331"). InnerVolumeSpecName "kube-api-access-hvlrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:30:03 crc kubenswrapper[4725]: I1202 13:30:03.035837 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvlrx\" (UniqueName: \"kubernetes.io/projected/5c51f8b2-f00f-47e4-92ae-000fac82e331-kube-api-access-hvlrx\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:03 crc kubenswrapper[4725]: I1202 13:30:03.035874 4725 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5c51f8b2-f00f-47e4-92ae-000fac82e331-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:03 crc kubenswrapper[4725]: I1202 13:30:03.540777 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" event={"ID":"5c51f8b2-f00f-47e4-92ae-000fac82e331","Type":"ContainerDied","Data":"38c9d5eae8e821065072584fb5dd9904dac8b19512f6b3ebbbd6f5c2e0978525"} Dec 02 13:30:03 crc kubenswrapper[4725]: I1202 13:30:03.541099 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38c9d5eae8e821065072584fb5dd9904dac8b19512f6b3ebbbd6f5c2e0978525" Dec 02 13:30:03 crc kubenswrapper[4725]: I1202 13:30:03.540827 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz" Dec 02 13:30:06 crc kubenswrapper[4725]: I1202 13:30:06.269147 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:30:06 crc kubenswrapper[4725]: E1202 13:30:06.270149 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:30:19 crc kubenswrapper[4725]: I1202 13:30:19.276122 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:30:19 crc kubenswrapper[4725]: E1202 13:30:19.276922 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.656374 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gdwqj"] Dec 02 13:30:23 crc kubenswrapper[4725]: E1202 13:30:23.657847 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c51f8b2-f00f-47e4-92ae-000fac82e331" containerName="collect-profiles" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.657863 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c51f8b2-f00f-47e4-92ae-000fac82e331" containerName="collect-profiles" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.658138 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c51f8b2-f00f-47e4-92ae-000fac82e331" containerName="collect-profiles" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.659982 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.667920 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gdwqj"] Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.860251 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-catalog-content\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.860654 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-875b8\" (UniqueName: \"kubernetes.io/projected/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-kube-api-access-875b8\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.861162 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-utilities\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.964799 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-catalog-content\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.964961 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-875b8\" (UniqueName: \"kubernetes.io/projected/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-kube-api-access-875b8\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.965044 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-utilities\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.965558 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-catalog-content\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.965592 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-utilities\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:23 crc kubenswrapper[4725]: I1202 13:30:23.993526 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-875b8\" (UniqueName: \"kubernetes.io/projected/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-kube-api-access-875b8\") pod \"certified-operators-gdwqj\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:24 crc kubenswrapper[4725]: I1202 13:30:24.293938 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:24 crc kubenswrapper[4725]: I1202 13:30:24.830451 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gdwqj"] Dec 02 13:30:25 crc kubenswrapper[4725]: I1202 13:30:25.770124 4725 generic.go:334] "Generic (PLEG): container finished" podID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerID="93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c" exitCode=0 Dec 02 13:30:25 crc kubenswrapper[4725]: I1202 13:30:25.770361 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdwqj" event={"ID":"febfddaa-dbe6-4d0f-bf28-de7b823f78a6","Type":"ContainerDied","Data":"93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c"} Dec 02 13:30:25 crc kubenswrapper[4725]: I1202 13:30:25.770713 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdwqj" event={"ID":"febfddaa-dbe6-4d0f-bf28-de7b823f78a6","Type":"ContainerStarted","Data":"ae68e475b14bb39f6fc34ca2fd81706ecdcf9a64d00ecaa28b595d96424cd6b2"} Dec 02 13:30:25 crc kubenswrapper[4725]: I1202 13:30:25.773803 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:30:28 crc kubenswrapper[4725]: I1202 13:30:28.809048 4725 generic.go:334] "Generic (PLEG): container finished" podID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerID="73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27" exitCode=0 Dec 02 13:30:28 crc kubenswrapper[4725]: I1202 13:30:28.809714 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdwqj" event={"ID":"febfddaa-dbe6-4d0f-bf28-de7b823f78a6","Type":"ContainerDied","Data":"73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27"} Dec 02 13:30:30 crc kubenswrapper[4725]: I1202 13:30:30.267971 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:30:30 crc kubenswrapper[4725]: E1202 13:30:30.268590 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:30:32 crc kubenswrapper[4725]: I1202 13:30:32.870751 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdwqj" event={"ID":"febfddaa-dbe6-4d0f-bf28-de7b823f78a6","Type":"ContainerStarted","Data":"d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27"} Dec 02 13:30:32 crc kubenswrapper[4725]: I1202 13:30:32.892852 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gdwqj" podStartSLOduration=3.470794942 podStartE2EDuration="9.892834097s" podCreationTimestamp="2025-12-02 13:30:23 +0000 UTC" firstStartedPulling="2025-12-02 13:30:25.773516456 +0000 UTC m=+1556.730158151" lastFinishedPulling="2025-12-02 13:30:32.195555591 +0000 UTC m=+1563.152197306" observedRunningTime="2025-12-02 13:30:32.890760185 +0000 UTC m=+1563.847401880" watchObservedRunningTime="2025-12-02 13:30:32.892834097 +0000 UTC m=+1563.849475792" Dec 02 13:30:34 crc kubenswrapper[4725]: I1202 13:30:34.294893 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:34 crc kubenswrapper[4725]: I1202 13:30:34.295847 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:34 crc kubenswrapper[4725]: I1202 13:30:34.345344 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:42 crc kubenswrapper[4725]: I1202 13:30:42.967525 4725 generic.go:334] "Generic (PLEG): container finished" podID="3ab59324-3446-4602-af02-e9094f220f34" containerID="28b80448f9548a67732564e76303892c724cab4fe81ba0a4325582f91da04604" exitCode=0 Dec 02 13:30:42 crc kubenswrapper[4725]: I1202 13:30:42.967610 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" event={"ID":"3ab59324-3446-4602-af02-e9094f220f34","Type":"ContainerDied","Data":"28b80448f9548a67732564e76303892c724cab4fe81ba0a4325582f91da04604"} Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.355443 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.428072 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gdwqj"] Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.526593 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.673700 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-bootstrap-combined-ca-bundle\") pod \"3ab59324-3446-4602-af02-e9094f220f34\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.674059 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-inventory\") pod \"3ab59324-3446-4602-af02-e9094f220f34\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.674305 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcdll\" (UniqueName: \"kubernetes.io/projected/3ab59324-3446-4602-af02-e9094f220f34-kube-api-access-xcdll\") pod \"3ab59324-3446-4602-af02-e9094f220f34\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.674354 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-ssh-key\") pod \"3ab59324-3446-4602-af02-e9094f220f34\" (UID: \"3ab59324-3446-4602-af02-e9094f220f34\") " Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.682971 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab59324-3446-4602-af02-e9094f220f34-kube-api-access-xcdll" (OuterVolumeSpecName: "kube-api-access-xcdll") pod "3ab59324-3446-4602-af02-e9094f220f34" (UID: "3ab59324-3446-4602-af02-e9094f220f34"). InnerVolumeSpecName "kube-api-access-xcdll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.683314 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3ab59324-3446-4602-af02-e9094f220f34" (UID: "3ab59324-3446-4602-af02-e9094f220f34"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.717643 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3ab59324-3446-4602-af02-e9094f220f34" (UID: "3ab59324-3446-4602-af02-e9094f220f34"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.727190 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-inventory" (OuterVolumeSpecName: "inventory") pod "3ab59324-3446-4602-af02-e9094f220f34" (UID: "3ab59324-3446-4602-af02-e9094f220f34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.777640 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcdll\" (UniqueName: \"kubernetes.io/projected/3ab59324-3446-4602-af02-e9094f220f34-kube-api-access-xcdll\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.777687 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.777700 4725 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.777711 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ab59324-3446-4602-af02-e9094f220f34-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.992546 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.992574 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr" event={"ID":"3ab59324-3446-4602-af02-e9094f220f34","Type":"ContainerDied","Data":"8ba76f9ebaa74940c0ac64c6e036470e01ba1b3d4b855203eecf60dc8813c74b"} Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.992621 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ba76f9ebaa74940c0ac64c6e036470e01ba1b3d4b855203eecf60dc8813c74b" Dec 02 13:30:44 crc kubenswrapper[4725]: I1202 13:30:44.992675 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gdwqj" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="registry-server" containerID="cri-o://d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27" gracePeriod=2 Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.105626 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg"] Dec 02 13:30:45 crc kubenswrapper[4725]: E1202 13:30:45.106487 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ab59324-3446-4602-af02-e9094f220f34" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.106518 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ab59324-3446-4602-af02-e9094f220f34" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.106801 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ab59324-3446-4602-af02-e9094f220f34" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.107947 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.111105 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.111352 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.112170 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.114041 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.119315 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg"] Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.184844 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.184921 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.185488 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpztt\" (UniqueName: \"kubernetes.io/projected/022c7b77-3848-4ca6-9382-64eb9762bf40-kube-api-access-qpztt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.292186 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.292272 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.292382 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpztt\" (UniqueName: \"kubernetes.io/projected/022c7b77-3848-4ca6-9382-64eb9762bf40-kube-api-access-qpztt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.302517 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.302972 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.311163 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpztt\" (UniqueName: \"kubernetes.io/projected/022c7b77-3848-4ca6-9382-64eb9762bf40-kube-api-access-qpztt\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.471746 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.545168 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.708618 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-utilities\") pod \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.708842 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-875b8\" (UniqueName: \"kubernetes.io/projected/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-kube-api-access-875b8\") pod \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.708873 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-catalog-content\") pod \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\" (UID: \"febfddaa-dbe6-4d0f-bf28-de7b823f78a6\") " Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.709916 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-utilities" (OuterVolumeSpecName: "utilities") pod "febfddaa-dbe6-4d0f-bf28-de7b823f78a6" (UID: "febfddaa-dbe6-4d0f-bf28-de7b823f78a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.715300 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-kube-api-access-875b8" (OuterVolumeSpecName: "kube-api-access-875b8") pod "febfddaa-dbe6-4d0f-bf28-de7b823f78a6" (UID: "febfddaa-dbe6-4d0f-bf28-de7b823f78a6"). InnerVolumeSpecName "kube-api-access-875b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.765354 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "febfddaa-dbe6-4d0f-bf28-de7b823f78a6" (UID: "febfddaa-dbe6-4d0f-bf28-de7b823f78a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.812683 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.812855 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-875b8\" (UniqueName: \"kubernetes.io/projected/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-kube-api-access-875b8\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:45 crc kubenswrapper[4725]: I1202 13:30:45.813057 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/febfddaa-dbe6-4d0f-bf28-de7b823f78a6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.005888 4725 generic.go:334] "Generic (PLEG): container finished" podID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerID="d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27" exitCode=0 Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.005949 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gdwqj" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.005965 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdwqj" event={"ID":"febfddaa-dbe6-4d0f-bf28-de7b823f78a6","Type":"ContainerDied","Data":"d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27"} Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.006026 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gdwqj" event={"ID":"febfddaa-dbe6-4d0f-bf28-de7b823f78a6","Type":"ContainerDied","Data":"ae68e475b14bb39f6fc34ca2fd81706ecdcf9a64d00ecaa28b595d96424cd6b2"} Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.006053 4725 scope.go:117] "RemoveContainer" containerID="d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.036073 4725 scope.go:117] "RemoveContainer" containerID="73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.052710 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gdwqj"] Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.064330 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gdwqj"] Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.066126 4725 scope.go:117] "RemoveContainer" containerID="93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.094310 4725 scope.go:117] "RemoveContainer" containerID="d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27" Dec 02 13:30:46 crc kubenswrapper[4725]: E1202 13:30:46.094864 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27\": container with ID starting with d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27 not found: ID does not exist" containerID="d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.094910 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27"} err="failed to get container status \"d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27\": rpc error: code = NotFound desc = could not find container \"d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27\": container with ID starting with d8ed1618d986514684a1876413548dd8b5b3ec140751ce975c55e347cb629b27 not found: ID does not exist" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.094947 4725 scope.go:117] "RemoveContainer" containerID="73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27" Dec 02 13:30:46 crc kubenswrapper[4725]: E1202 13:30:46.095299 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27\": container with ID starting with 73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27 not found: ID does not exist" containerID="73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.095350 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27"} err="failed to get container status \"73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27\": rpc error: code = NotFound desc = could not find container \"73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27\": container with ID starting with 73f01d0670b85f862059725e1f1f17c3cb4ed94e26ecd7f80b009993335e3e27 not found: ID does not exist" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.095388 4725 scope.go:117] "RemoveContainer" containerID="93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c" Dec 02 13:30:46 crc kubenswrapper[4725]: E1202 13:30:46.096220 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c\": container with ID starting with 93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c not found: ID does not exist" containerID="93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.096260 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c"} err="failed to get container status \"93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c\": rpc error: code = NotFound desc = could not find container \"93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c\": container with ID starting with 93b6f63026efe0f4db155341a4f3f43ca4f9924ca1d3d74be33780828799380c not found: ID does not exist" Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.133221 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg"] Dec 02 13:30:46 crc kubenswrapper[4725]: I1202 13:30:46.269044 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:30:46 crc kubenswrapper[4725]: E1202 13:30:46.269350 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:30:47 crc kubenswrapper[4725]: I1202 13:30:47.019844 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" event={"ID":"022c7b77-3848-4ca6-9382-64eb9762bf40","Type":"ContainerStarted","Data":"36e7d14eebd5eae3e9e2565a5cb9e90ff76deb7e7727ed86d8c7a1721457c2cf"} Dec 02 13:30:47 crc kubenswrapper[4725]: I1202 13:30:47.282926 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" path="/var/lib/kubelet/pods/febfddaa-dbe6-4d0f-bf28-de7b823f78a6/volumes" Dec 02 13:30:48 crc kubenswrapper[4725]: I1202 13:30:48.034762 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" event={"ID":"022c7b77-3848-4ca6-9382-64eb9762bf40","Type":"ContainerStarted","Data":"0a7cb3f055dc5e664ca5992d3e7a4ec7adde4c357d7a2d4cdb28708614a45d21"} Dec 02 13:30:48 crc kubenswrapper[4725]: I1202 13:30:48.065581 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" podStartSLOduration=2.279315689 podStartE2EDuration="3.065547577s" podCreationTimestamp="2025-12-02 13:30:45 +0000 UTC" firstStartedPulling="2025-12-02 13:30:46.137686389 +0000 UTC m=+1577.094328084" lastFinishedPulling="2025-12-02 13:30:46.923918277 +0000 UTC m=+1577.880559972" observedRunningTime="2025-12-02 13:30:48.04930818 +0000 UTC m=+1579.005949875" watchObservedRunningTime="2025-12-02 13:30:48.065547577 +0000 UTC m=+1579.022189272" Dec 02 13:30:57 crc kubenswrapper[4725]: I1202 13:30:57.268678 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:30:57 crc kubenswrapper[4725]: E1202 13:30:57.269493 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.045040 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-f6d7c"] Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.054864 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-81d8-account-create-update-tvqbp"] Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.070298 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-l9fs7"] Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.078523 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-f6d7c"] Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.086794 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-81d8-account-create-update-tvqbp"] Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.095870 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-l9fs7"] Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.280440 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5" path="/var/lib/kubelet/pods/90ffa1aa-9afc-47dd-a3c0-cb1358a68ad5/volumes" Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.281123 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c55f5277-1f0c-450f-8d58-54d808005c47" path="/var/lib/kubelet/pods/c55f5277-1f0c-450f-8d58-54d808005c47/volumes" Dec 02 13:31:05 crc kubenswrapper[4725]: I1202 13:31:05.281666 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf90bb11-ddb2-4433-82d3-b168999cc158" path="/var/lib/kubelet/pods/cf90bb11-ddb2-4433-82d3-b168999cc158/volumes" Dec 02 13:31:06 crc kubenswrapper[4725]: I1202 13:31:06.033441 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4596-account-create-update-nfdpv"] Dec 02 13:31:06 crc kubenswrapper[4725]: I1202 13:31:06.043276 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4596-account-create-update-nfdpv"] Dec 02 13:31:07 crc kubenswrapper[4725]: I1202 13:31:07.279725 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9793f192-d30b-47f0-a0e0-b31c07946211" path="/var/lib/kubelet/pods/9793f192-d30b-47f0-a0e0-b31c07946211/volumes" Dec 02 13:31:10 crc kubenswrapper[4725]: I1202 13:31:10.268910 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:31:10 crc kubenswrapper[4725]: E1202 13:31:10.271066 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:31:11 crc kubenswrapper[4725]: I1202 13:31:11.046006 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a1c9-account-create-update-jzlj6"] Dec 02 13:31:11 crc kubenswrapper[4725]: I1202 13:31:11.055873 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-vd2kp"] Dec 02 13:31:11 crc kubenswrapper[4725]: I1202 13:31:11.067256 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-vd2kp"] Dec 02 13:31:11 crc kubenswrapper[4725]: I1202 13:31:11.076600 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a1c9-account-create-update-jzlj6"] Dec 02 13:31:11 crc kubenswrapper[4725]: I1202 13:31:11.279641 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ffbb6e-2974-43d3-aef7-6ff8b1a62aab" path="/var/lib/kubelet/pods/16ffbb6e-2974-43d3-aef7-6ff8b1a62aab/volumes" Dec 02 13:31:11 crc kubenswrapper[4725]: I1202 13:31:11.280319 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f145bf96-9032-4d92-8aac-9ebe20153fca" path="/var/lib/kubelet/pods/f145bf96-9032-4d92-8aac-9ebe20153fca/volumes" Dec 02 13:31:22 crc kubenswrapper[4725]: I1202 13:31:22.268863 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:31:22 crc kubenswrapper[4725]: E1202 13:31:22.269925 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:31:35 crc kubenswrapper[4725]: I1202 13:31:35.268947 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:31:35 crc kubenswrapper[4725]: E1202 13:31:35.270038 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.523158 4725 scope.go:117] "RemoveContainer" containerID="44ed2cad4986e9f3fbaa33193b8751d418779348fd30be9177344d9cc05b684e" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.551825 4725 scope.go:117] "RemoveContainer" containerID="835ac8643da173853954018e41884fc776adee8dc9425aeb33fda07cf9f77ddf" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.574937 4725 scope.go:117] "RemoveContainer" containerID="576ae5c7c4c6166ce9e548c491b67e056fc5d7ca1b77bee37ff5c5b83cfe8789" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.631539 4725 scope.go:117] "RemoveContainer" containerID="18e29ee21f03cbf03585031abc2d12a9b72e9ccc532e1d82569a0c8dbba73d37" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.692570 4725 scope.go:117] "RemoveContainer" containerID="940a93f04e76a7ec1949b2d4e141548dc960bf004222efffd4d2475fc5a64529" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.734007 4725 scope.go:117] "RemoveContainer" containerID="c5e44a0b1ab0250d3acd27708fb2e8ece33ca14d37ac328b836021156ed7ca9b" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.775361 4725 scope.go:117] "RemoveContainer" containerID="d98e99acffb691b6d57cf4fadb743c9f0b12e32adc9fb5dd0dc66117bfb3e57b" Dec 02 13:31:47 crc kubenswrapper[4725]: I1202 13:31:47.803714 4725 scope.go:117] "RemoveContainer" containerID="82af7c2f7a6a510d4d8c7d20856cc39062a65d8de1b75c7144e34c99c62b62f8" Dec 02 13:31:49 crc kubenswrapper[4725]: I1202 13:31:49.275222 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:31:49 crc kubenswrapper[4725]: E1202 13:31:49.275788 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.083097 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e6bf-account-create-update-hhhqj"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.098190 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-5cvgg"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.113541 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tghwc"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.131675 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e6bf-account-create-update-hhhqj"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.142686 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8bl2c"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.152041 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-5cvgg"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.161874 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3f8e-account-create-update-wcchj"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.171440 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tghwc"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.181545 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8bl2c"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.191594 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3f8e-account-create-update-wcchj"] Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.282014 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019" path="/var/lib/kubelet/pods/10ca0cc7-f32f-40e8-9ce9-5a0d0efa1019/volumes" Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.283258 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8" path="/var/lib/kubelet/pods/1cddd3b3-e14f-4dba-a7f9-7ac6ae0bafa8/volumes" Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.284017 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3607fd3f-e379-42ee-b778-5d1540371a35" path="/var/lib/kubelet/pods/3607fd3f-e379-42ee-b778-5d1540371a35/volumes" Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.284628 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7" path="/var/lib/kubelet/pods/5ae7992f-1df6-4ed8-bd47-c99ff8b7ddd7/volumes" Dec 02 13:31:55 crc kubenswrapper[4725]: I1202 13:31:55.286607 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3946847-5d4a-4fcc-acac-7a49400db5b2" path="/var/lib/kubelet/pods/b3946847-5d4a-4fcc-acac-7a49400db5b2/volumes" Dec 02 13:31:56 crc kubenswrapper[4725]: I1202 13:31:56.034191 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3a9a-account-create-update-wkmbs"] Dec 02 13:31:56 crc kubenswrapper[4725]: I1202 13:31:56.044418 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3a9a-account-create-update-wkmbs"] Dec 02 13:31:57 crc kubenswrapper[4725]: I1202 13:31:57.280986 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5" path="/var/lib/kubelet/pods/1de8e7c6-e7c2-4212-bd1f-26fb976fc3d5/volumes" Dec 02 13:32:03 crc kubenswrapper[4725]: I1202 13:32:03.043800 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-7d27g"] Dec 02 13:32:03 crc kubenswrapper[4725]: I1202 13:32:03.058838 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-7d27g"] Dec 02 13:32:03 crc kubenswrapper[4725]: I1202 13:32:03.281404 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154d76e0-4eac-4a29-8f3a-5aadb8762c6b" path="/var/lib/kubelet/pods/154d76e0-4eac-4a29-8f3a-5aadb8762c6b/volumes" Dec 02 13:32:04 crc kubenswrapper[4725]: I1202 13:32:04.268827 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:32:04 crc kubenswrapper[4725]: E1202 13:32:04.269098 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:32:05 crc kubenswrapper[4725]: I1202 13:32:05.035994 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zmwkf"] Dec 02 13:32:05 crc kubenswrapper[4725]: I1202 13:32:05.044145 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zmwkf"] Dec 02 13:32:05 crc kubenswrapper[4725]: I1202 13:32:05.279021 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c76d8c86-54fe-4bb2-9590-76e2a56d8659" path="/var/lib/kubelet/pods/c76d8c86-54fe-4bb2-9590-76e2a56d8659/volumes" Dec 02 13:32:15 crc kubenswrapper[4725]: I1202 13:32:15.269639 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:32:15 crc kubenswrapper[4725]: E1202 13:32:15.270421 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:32:30 crc kubenswrapper[4725]: I1202 13:32:30.267927 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:32:30 crc kubenswrapper[4725]: E1202 13:32:30.268844 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:32:41 crc kubenswrapper[4725]: I1202 13:32:41.046045 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-lx9dr"] Dec 02 13:32:41 crc kubenswrapper[4725]: I1202 13:32:41.054401 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-lx9dr"] Dec 02 13:32:41 crc kubenswrapper[4725]: I1202 13:32:41.281775 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ed1edd-ee3b-48a5-b4a6-539e780a2e76" path="/var/lib/kubelet/pods/b3ed1edd-ee3b-48a5-b4a6-539e780a2e76/volumes" Dec 02 13:32:43 crc kubenswrapper[4725]: I1202 13:32:43.029957 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-psww9"] Dec 02 13:32:43 crc kubenswrapper[4725]: I1202 13:32:43.041120 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-psww9"] Dec 02 13:32:43 crc kubenswrapper[4725]: I1202 13:32:43.206822 4725 generic.go:334] "Generic (PLEG): container finished" podID="022c7b77-3848-4ca6-9382-64eb9762bf40" containerID="0a7cb3f055dc5e664ca5992d3e7a4ec7adde4c357d7a2d4cdb28708614a45d21" exitCode=0 Dec 02 13:32:43 crc kubenswrapper[4725]: I1202 13:32:43.206922 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" event={"ID":"022c7b77-3848-4ca6-9382-64eb9762bf40","Type":"ContainerDied","Data":"0a7cb3f055dc5e664ca5992d3e7a4ec7adde4c357d7a2d4cdb28708614a45d21"} Dec 02 13:32:43 crc kubenswrapper[4725]: I1202 13:32:43.278393 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e" path="/var/lib/kubelet/pods/b9dfb71a-3730-4e3a-b85d-fd3b8bfacb6e/volumes" Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.848661 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.930659 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-ssh-key\") pod \"022c7b77-3848-4ca6-9382-64eb9762bf40\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.930718 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-inventory\") pod \"022c7b77-3848-4ca6-9382-64eb9762bf40\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.930860 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpztt\" (UniqueName: \"kubernetes.io/projected/022c7b77-3848-4ca6-9382-64eb9762bf40-kube-api-access-qpztt\") pod \"022c7b77-3848-4ca6-9382-64eb9762bf40\" (UID: \"022c7b77-3848-4ca6-9382-64eb9762bf40\") " Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.936349 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/022c7b77-3848-4ca6-9382-64eb9762bf40-kube-api-access-qpztt" (OuterVolumeSpecName: "kube-api-access-qpztt") pod "022c7b77-3848-4ca6-9382-64eb9762bf40" (UID: "022c7b77-3848-4ca6-9382-64eb9762bf40"). InnerVolumeSpecName "kube-api-access-qpztt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.959841 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "022c7b77-3848-4ca6-9382-64eb9762bf40" (UID: "022c7b77-3848-4ca6-9382-64eb9762bf40"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:32:44 crc kubenswrapper[4725]: I1202 13:32:44.960075 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-inventory" (OuterVolumeSpecName: "inventory") pod "022c7b77-3848-4ca6-9382-64eb9762bf40" (UID: "022c7b77-3848-4ca6-9382-64eb9762bf40"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.033049 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.033090 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022c7b77-3848-4ca6-9382-64eb9762bf40-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.033105 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpztt\" (UniqueName: \"kubernetes.io/projected/022c7b77-3848-4ca6-9382-64eb9762bf40-kube-api-access-qpztt\") on node \"crc\" DevicePath \"\"" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.228658 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.229761 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg" event={"ID":"022c7b77-3848-4ca6-9382-64eb9762bf40","Type":"ContainerDied","Data":"36e7d14eebd5eae3e9e2565a5cb9e90ff76deb7e7727ed86d8c7a1721457c2cf"} Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.229820 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36e7d14eebd5eae3e9e2565a5cb9e90ff76deb7e7727ed86d8c7a1721457c2cf" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.267913 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:32:45 crc kubenswrapper[4725]: E1202 13:32:45.268384 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.315435 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc"] Dec 02 13:32:45 crc kubenswrapper[4725]: E1202 13:32:45.316064 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="registry-server" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.316091 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="registry-server" Dec 02 13:32:45 crc kubenswrapper[4725]: E1202 13:32:45.316131 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="extract-content" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.316140 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="extract-content" Dec 02 13:32:45 crc kubenswrapper[4725]: E1202 13:32:45.316158 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="022c7b77-3848-4ca6-9382-64eb9762bf40" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.316168 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="022c7b77-3848-4ca6-9382-64eb9762bf40" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 13:32:45 crc kubenswrapper[4725]: E1202 13:32:45.316178 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="extract-utilities" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.316185 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="extract-utilities" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.316425 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="022c7b77-3848-4ca6-9382-64eb9762bf40" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.316451 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="febfddaa-dbe6-4d0f-bf28-de7b823f78a6" containerName="registry-server" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.317287 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.319921 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.320289 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.321084 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.326326 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.327290 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc"] Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.343025 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.343134 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m94mc\" (UniqueName: \"kubernetes.io/projected/711edc3c-a714-4e16-a7b8-6f61529f69d7-kube-api-access-m94mc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.343582 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.444981 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.445029 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.445069 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m94mc\" (UniqueName: \"kubernetes.io/projected/711edc3c-a714-4e16-a7b8-6f61529f69d7-kube-api-access-m94mc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.450155 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.460579 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.469888 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m94mc\" (UniqueName: \"kubernetes.io/projected/711edc3c-a714-4e16-a7b8-6f61529f69d7-kube-api-access-m94mc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:45 crc kubenswrapper[4725]: I1202 13:32:45.645061 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:32:46 crc kubenswrapper[4725]: I1202 13:32:46.160446 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc"] Dec 02 13:32:46 crc kubenswrapper[4725]: I1202 13:32:46.239401 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" event={"ID":"711edc3c-a714-4e16-a7b8-6f61529f69d7","Type":"ContainerStarted","Data":"f82ce3fdc13bfa6b99aeb1e362be0c1202b5c3df420c4f0a5163b9674593a3ff"} Dec 02 13:32:47 crc kubenswrapper[4725]: I1202 13:32:47.250854 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" event={"ID":"711edc3c-a714-4e16-a7b8-6f61529f69d7","Type":"ContainerStarted","Data":"8bc928c48164ca9c9615f3c503b9487456e41191bdacbd6cc6425f643d6ef3f0"} Dec 02 13:32:47 crc kubenswrapper[4725]: I1202 13:32:47.266745 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" podStartSLOduration=1.664110948 podStartE2EDuration="2.266719484s" podCreationTimestamp="2025-12-02 13:32:45 +0000 UTC" firstStartedPulling="2025-12-02 13:32:46.168833795 +0000 UTC m=+1697.125475490" lastFinishedPulling="2025-12-02 13:32:46.771442331 +0000 UTC m=+1697.728084026" observedRunningTime="2025-12-02 13:32:47.265401731 +0000 UTC m=+1698.222043436" watchObservedRunningTime="2025-12-02 13:32:47.266719484 +0000 UTC m=+1698.223361179" Dec 02 13:32:47 crc kubenswrapper[4725]: I1202 13:32:47.956069 4725 scope.go:117] "RemoveContainer" containerID="a4e89f6e9c6fa2fe3742e3ceaa282dd38e3e0396a9f1e04f6b7d623789005d36" Dec 02 13:32:47 crc kubenswrapper[4725]: I1202 13:32:47.979171 4725 scope.go:117] "RemoveContainer" containerID="39ea52ea6285c01b3b519293f937ef8f9d19b4e72bc46fa935f7e64f87e1b08d" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.048138 4725 scope.go:117] "RemoveContainer" containerID="3f90d3ab25f62602b6bc226bb3cf96e98918e5370b9e3e7a2bb566c8fb637d61" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.088336 4725 scope.go:117] "RemoveContainer" containerID="6a99a7a44b1ca76d56407724b6adb597b9c0ff0392cc4225b7f17397752169c4" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.167267 4725 scope.go:117] "RemoveContainer" containerID="28b54163a4f9afb5c4fb334e58ae4e155c4241896fe8be52b62a553a49de9d86" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.193839 4725 scope.go:117] "RemoveContainer" containerID="1b967543fe877f2d2628c57a87a3c6aeccf0ccb6407d44af3ec41221292a1f16" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.239809 4725 scope.go:117] "RemoveContainer" containerID="0f9f97eecdfc4c2d1cca68aba5c072eb9b014f2a535efb2d918522dbc660ccf4" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.260885 4725 scope.go:117] "RemoveContainer" containerID="1e9f5e0868aa10e463afac90fa39f6c8ae20b9ffa2bc874ff0e970d7f0874046" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.285528 4725 scope.go:117] "RemoveContainer" containerID="b85444a496bd38286c3f1aecb2e9905fc40447f2bc1e19cd6485fb817cad8caa" Dec 02 13:32:48 crc kubenswrapper[4725]: I1202 13:32:48.316709 4725 scope.go:117] "RemoveContainer" containerID="7937fc5aee77e9ecb0e9825cde9bfc317794afb497fb367d1c038e131d473544" Dec 02 13:32:53 crc kubenswrapper[4725]: I1202 13:32:53.041325 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kzkrh"] Dec 02 13:32:53 crc kubenswrapper[4725]: I1202 13:32:53.049968 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kzkrh"] Dec 02 13:32:53 crc kubenswrapper[4725]: I1202 13:32:53.278896 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3128bc10-71dd-46f4-8e71-787260ede3d0" path="/var/lib/kubelet/pods/3128bc10-71dd-46f4-8e71-787260ede3d0/volumes" Dec 02 13:32:56 crc kubenswrapper[4725]: I1202 13:32:56.268545 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:32:56 crc kubenswrapper[4725]: E1202 13:32:56.269179 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.038870 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-9blbh"] Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.051576 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-9blbh"] Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.077227 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-pdwcp"] Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.098786 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-pdwcp"] Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.274639 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:33:09 crc kubenswrapper[4725]: E1202 13:33:09.274895 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.279161 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11048035-cfbd-4e03-8427-83e0d2a91a63" path="/var/lib/kubelet/pods/11048035-cfbd-4e03-8427-83e0d2a91a63/volumes" Dec 02 13:33:09 crc kubenswrapper[4725]: I1202 13:33:09.280201 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52677bda-0bda-4d69-af01-f6ce198582f6" path="/var/lib/kubelet/pods/52677bda-0bda-4d69-af01-f6ce198582f6/volumes" Dec 02 13:33:24 crc kubenswrapper[4725]: I1202 13:33:24.269094 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:33:24 crc kubenswrapper[4725]: E1202 13:33:24.269848 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:33:38 crc kubenswrapper[4725]: I1202 13:33:38.268848 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:33:38 crc kubenswrapper[4725]: E1202 13:33:38.269654 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:33:48 crc kubenswrapper[4725]: I1202 13:33:48.521510 4725 scope.go:117] "RemoveContainer" containerID="2f46d8863cae3b7595faf9b2e979579633b73b63008a54b957ca0ce9dc6bee18" Dec 02 13:33:48 crc kubenswrapper[4725]: I1202 13:33:48.548755 4725 scope.go:117] "RemoveContainer" containerID="ca608bf96ccf5468c525dacc3104084d32b250d892bac9d8deb4b016b1735347" Dec 02 13:33:48 crc kubenswrapper[4725]: I1202 13:33:48.625492 4725 scope.go:117] "RemoveContainer" containerID="defad413a0b029b7aa4a113098e7b7a85fd86214d0d36a873dc97ae4d29e840d" Dec 02 13:33:48 crc kubenswrapper[4725]: I1202 13:33:48.696363 4725 scope.go:117] "RemoveContainer" containerID="bf9a62d00dd6f0ceb4906e0f581db0961eb660e16074e763ad0730925d8188ee" Dec 02 13:33:48 crc kubenswrapper[4725]: I1202 13:33:48.735309 4725 scope.go:117] "RemoveContainer" containerID="68d371efd80ed263528e6349b6ee85336db6e7d51973a405585eb2a9f5a2e8e3" Dec 02 13:33:48 crc kubenswrapper[4725]: I1202 13:33:48.762799 4725 scope.go:117] "RemoveContainer" containerID="40b77c20241ca086c6d9446589a4b907e013982c8ff95fd853ad2cec508cfd4c" Dec 02 13:33:49 crc kubenswrapper[4725]: I1202 13:33:49.285041 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:33:49 crc kubenswrapper[4725]: E1202 13:33:49.286080 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.045814 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-qrnph"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.055437 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-79dd-account-create-update-h92mv"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.064667 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b3bb-account-create-update-hjg7w"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.075132 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xkn8g"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.083975 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-bmtlh"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.091187 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-80a5-account-create-update-4pzhq"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.098219 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-79dd-account-create-update-h92mv"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.105166 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-qrnph"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.113647 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-bmtlh"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.122090 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xkn8g"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.129501 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-80a5-account-create-update-4pzhq"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.136299 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b3bb-account-create-update-hjg7w"] Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.279360 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a9ce041-5d7a-4a5f-be52-fdfae2f06a53" path="/var/lib/kubelet/pods/1a9ce041-5d7a-4a5f-be52-fdfae2f06a53/volumes" Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.280574 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e26ab0f-08f5-4712-a40d-ff08b563b752" path="/var/lib/kubelet/pods/1e26ab0f-08f5-4712-a40d-ff08b563b752/volumes" Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.281519 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a1f9729-56c5-4955-86b3-68f05bb15396" path="/var/lib/kubelet/pods/2a1f9729-56c5-4955-86b3-68f05bb15396/volumes" Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.282574 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a0d2b3-489c-4160-a7c0-7570d7baca42" path="/var/lib/kubelet/pods/46a0d2b3-489c-4160-a7c0-7570d7baca42/volumes" Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.284329 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0d25cc-d587-4265-b6d3-373ea0e46728" path="/var/lib/kubelet/pods/5c0d25cc-d587-4265-b6d3-373ea0e46728/volumes" Dec 02 13:33:55 crc kubenswrapper[4725]: I1202 13:33:55.285215 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d35ce79f-44d1-4180-a786-32dd8df46b76" path="/var/lib/kubelet/pods/d35ce79f-44d1-4180-a786-32dd8df46b76/volumes" Dec 02 13:33:59 crc kubenswrapper[4725]: I1202 13:33:59.938797 4725 generic.go:334] "Generic (PLEG): container finished" podID="711edc3c-a714-4e16-a7b8-6f61529f69d7" containerID="8bc928c48164ca9c9615f3c503b9487456e41191bdacbd6cc6425f643d6ef3f0" exitCode=0 Dec 02 13:33:59 crc kubenswrapper[4725]: I1202 13:33:59.938926 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" event={"ID":"711edc3c-a714-4e16-a7b8-6f61529f69d7","Type":"ContainerDied","Data":"8bc928c48164ca9c9615f3c503b9487456e41191bdacbd6cc6425f643d6ef3f0"} Dec 02 13:34:00 crc kubenswrapper[4725]: I1202 13:34:00.268632 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:34:00 crc kubenswrapper[4725]: E1202 13:34:00.269032 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.390932 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.425572 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m94mc\" (UniqueName: \"kubernetes.io/projected/711edc3c-a714-4e16-a7b8-6f61529f69d7-kube-api-access-m94mc\") pod \"711edc3c-a714-4e16-a7b8-6f61529f69d7\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.425913 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-ssh-key\") pod \"711edc3c-a714-4e16-a7b8-6f61529f69d7\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.426016 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-inventory\") pod \"711edc3c-a714-4e16-a7b8-6f61529f69d7\" (UID: \"711edc3c-a714-4e16-a7b8-6f61529f69d7\") " Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.431787 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711edc3c-a714-4e16-a7b8-6f61529f69d7-kube-api-access-m94mc" (OuterVolumeSpecName: "kube-api-access-m94mc") pod "711edc3c-a714-4e16-a7b8-6f61529f69d7" (UID: "711edc3c-a714-4e16-a7b8-6f61529f69d7"). InnerVolumeSpecName "kube-api-access-m94mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.456301 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-inventory" (OuterVolumeSpecName: "inventory") pod "711edc3c-a714-4e16-a7b8-6f61529f69d7" (UID: "711edc3c-a714-4e16-a7b8-6f61529f69d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.458401 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "711edc3c-a714-4e16-a7b8-6f61529f69d7" (UID: "711edc3c-a714-4e16-a7b8-6f61529f69d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.528024 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m94mc\" (UniqueName: \"kubernetes.io/projected/711edc3c-a714-4e16-a7b8-6f61529f69d7-kube-api-access-m94mc\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.528060 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.528070 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711edc3c-a714-4e16-a7b8-6f61529f69d7-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.958136 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" event={"ID":"711edc3c-a714-4e16-a7b8-6f61529f69d7","Type":"ContainerDied","Data":"f82ce3fdc13bfa6b99aeb1e362be0c1202b5c3df420c4f0a5163b9674593a3ff"} Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.958179 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f82ce3fdc13bfa6b99aeb1e362be0c1202b5c3df420c4f0a5163b9674593a3ff" Dec 02 13:34:01 crc kubenswrapper[4725]: I1202 13:34:01.958210 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.034298 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld"] Dec 02 13:34:02 crc kubenswrapper[4725]: E1202 13:34:02.034855 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711edc3c-a714-4e16-a7b8-6f61529f69d7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.034881 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="711edc3c-a714-4e16-a7b8-6f61529f69d7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.035143 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="711edc3c-a714-4e16-a7b8-6f61529f69d7" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.036100 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.037607 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.037716 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.037863 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvq75\" (UniqueName: \"kubernetes.io/projected/4a420694-62a2-4fd1-aade-ebfaaf73590c-kube-api-access-nvq75\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.038575 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.038687 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.038989 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.039108 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.043936 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld"] Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.139158 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvq75\" (UniqueName: \"kubernetes.io/projected/4a420694-62a2-4fd1-aade-ebfaaf73590c-kube-api-access-nvq75\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.139285 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.139317 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.143229 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.143377 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.157184 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvq75\" (UniqueName: \"kubernetes.io/projected/4a420694-62a2-4fd1-aade-ebfaaf73590c-kube-api-access-nvq75\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-dgwld\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.356497 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.857004 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld"] Dec 02 13:34:02 crc kubenswrapper[4725]: W1202 13:34:02.865058 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a420694_62a2_4fd1_aade_ebfaaf73590c.slice/crio-a0ad00507ecb008e1813fab91a63007ec272c7092eb62fde4c9852ea15212ef5 WatchSource:0}: Error finding container a0ad00507ecb008e1813fab91a63007ec272c7092eb62fde4c9852ea15212ef5: Status 404 returned error can't find the container with id a0ad00507ecb008e1813fab91a63007ec272c7092eb62fde4c9852ea15212ef5 Dec 02 13:34:02 crc kubenswrapper[4725]: I1202 13:34:02.973539 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" event={"ID":"4a420694-62a2-4fd1-aade-ebfaaf73590c","Type":"ContainerStarted","Data":"a0ad00507ecb008e1813fab91a63007ec272c7092eb62fde4c9852ea15212ef5"} Dec 02 13:34:04 crc kubenswrapper[4725]: I1202 13:34:04.991950 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" event={"ID":"4a420694-62a2-4fd1-aade-ebfaaf73590c","Type":"ContainerStarted","Data":"bee6818529ff5438eb3b3b278fa73ec4a9bcdbf3ea96268db4d84cc7092f7d46"} Dec 02 13:34:05 crc kubenswrapper[4725]: I1202 13:34:05.010199 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" podStartSLOduration=2.142992099 podStartE2EDuration="3.010181051s" podCreationTimestamp="2025-12-02 13:34:02 +0000 UTC" firstStartedPulling="2025-12-02 13:34:02.867584386 +0000 UTC m=+1773.824226081" lastFinishedPulling="2025-12-02 13:34:03.734773338 +0000 UTC m=+1774.691415033" observedRunningTime="2025-12-02 13:34:05.01012331 +0000 UTC m=+1775.966765015" watchObservedRunningTime="2025-12-02 13:34:05.010181051 +0000 UTC m=+1775.966822746" Dec 02 13:34:09 crc kubenswrapper[4725]: I1202 13:34:09.029818 4725 generic.go:334] "Generic (PLEG): container finished" podID="4a420694-62a2-4fd1-aade-ebfaaf73590c" containerID="bee6818529ff5438eb3b3b278fa73ec4a9bcdbf3ea96268db4d84cc7092f7d46" exitCode=0 Dec 02 13:34:09 crc kubenswrapper[4725]: I1202 13:34:09.029887 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" event={"ID":"4a420694-62a2-4fd1-aade-ebfaaf73590c","Type":"ContainerDied","Data":"bee6818529ff5438eb3b3b278fa73ec4a9bcdbf3ea96268db4d84cc7092f7d46"} Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.489156 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.608697 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-inventory\") pod \"4a420694-62a2-4fd1-aade-ebfaaf73590c\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.608821 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-ssh-key\") pod \"4a420694-62a2-4fd1-aade-ebfaaf73590c\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.608953 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvq75\" (UniqueName: \"kubernetes.io/projected/4a420694-62a2-4fd1-aade-ebfaaf73590c-kube-api-access-nvq75\") pod \"4a420694-62a2-4fd1-aade-ebfaaf73590c\" (UID: \"4a420694-62a2-4fd1-aade-ebfaaf73590c\") " Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.616925 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a420694-62a2-4fd1-aade-ebfaaf73590c-kube-api-access-nvq75" (OuterVolumeSpecName: "kube-api-access-nvq75") pod "4a420694-62a2-4fd1-aade-ebfaaf73590c" (UID: "4a420694-62a2-4fd1-aade-ebfaaf73590c"). InnerVolumeSpecName "kube-api-access-nvq75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.640755 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-inventory" (OuterVolumeSpecName: "inventory") pod "4a420694-62a2-4fd1-aade-ebfaaf73590c" (UID: "4a420694-62a2-4fd1-aade-ebfaaf73590c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.649346 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a420694-62a2-4fd1-aade-ebfaaf73590c" (UID: "4a420694-62a2-4fd1-aade-ebfaaf73590c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.711637 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.711676 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a420694-62a2-4fd1-aade-ebfaaf73590c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:10 crc kubenswrapper[4725]: I1202 13:34:10.711687 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvq75\" (UniqueName: \"kubernetes.io/projected/4a420694-62a2-4fd1-aade-ebfaaf73590c-kube-api-access-nvq75\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.048400 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" event={"ID":"4a420694-62a2-4fd1-aade-ebfaaf73590c","Type":"ContainerDied","Data":"a0ad00507ecb008e1813fab91a63007ec272c7092eb62fde4c9852ea15212ef5"} Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.048447 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0ad00507ecb008e1813fab91a63007ec272c7092eb62fde4c9852ea15212ef5" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.048520 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-dgwld" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.120895 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs"] Dec 02 13:34:11 crc kubenswrapper[4725]: E1202 13:34:11.121312 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a420694-62a2-4fd1-aade-ebfaaf73590c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.121333 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a420694-62a2-4fd1-aade-ebfaaf73590c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.121635 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a420694-62a2-4fd1-aade-ebfaaf73590c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.122375 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.126045 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.126320 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.126483 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.128990 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.136711 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs"] Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.220819 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn725\" (UniqueName: \"kubernetes.io/projected/375b1a65-e012-4cfa-a08b-5d0b321be9d7-kube-api-access-pn725\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.220863 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.221066 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.323577 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn725\" (UniqueName: \"kubernetes.io/projected/375b1a65-e012-4cfa-a08b-5d0b321be9d7-kube-api-access-pn725\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.323646 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.323717 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.330087 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.333219 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.340278 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn725\" (UniqueName: \"kubernetes.io/projected/375b1a65-e012-4cfa-a08b-5d0b321be9d7-kube-api-access-pn725\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-bxpzs\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.442027 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:11 crc kubenswrapper[4725]: I1202 13:34:11.944834 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs"] Dec 02 13:34:12 crc kubenswrapper[4725]: I1202 13:34:12.059514 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" event={"ID":"375b1a65-e012-4cfa-a08b-5d0b321be9d7","Type":"ContainerStarted","Data":"27637c00d7bae4fda9edc3433f58505a0316dd40c236d0898f40355193918626"} Dec 02 13:34:13 crc kubenswrapper[4725]: I1202 13:34:13.276650 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:34:13 crc kubenswrapper[4725]: E1202 13:34:13.277393 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:34:14 crc kubenswrapper[4725]: I1202 13:34:14.076906 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" event={"ID":"375b1a65-e012-4cfa-a08b-5d0b321be9d7","Type":"ContainerStarted","Data":"2699d2b16a8b9dd024e73bc6668d4ff63607a8d33aa522ec5fb5bcbbfc1e51b7"} Dec 02 13:34:14 crc kubenswrapper[4725]: I1202 13:34:14.097365 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" podStartSLOduration=2.260610009 podStartE2EDuration="3.097344939s" podCreationTimestamp="2025-12-02 13:34:11 +0000 UTC" firstStartedPulling="2025-12-02 13:34:11.959313548 +0000 UTC m=+1782.915955243" lastFinishedPulling="2025-12-02 13:34:12.796048478 +0000 UTC m=+1783.752690173" observedRunningTime="2025-12-02 13:34:14.090440997 +0000 UTC m=+1785.047082682" watchObservedRunningTime="2025-12-02 13:34:14.097344939 +0000 UTC m=+1785.053986634" Dec 02 13:34:25 crc kubenswrapper[4725]: I1202 13:34:25.268994 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:34:25 crc kubenswrapper[4725]: E1202 13:34:25.269849 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:34:27 crc kubenswrapper[4725]: I1202 13:34:27.043748 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z6bgr"] Dec 02 13:34:27 crc kubenswrapper[4725]: I1202 13:34:27.054522 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z6bgr"] Dec 02 13:34:27 crc kubenswrapper[4725]: I1202 13:34:27.282834 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2796bcd6-70d6-4895-9a2b-9d0de21ffa5e" path="/var/lib/kubelet/pods/2796bcd6-70d6-4895-9a2b-9d0de21ffa5e/volumes" Dec 02 13:34:40 crc kubenswrapper[4725]: I1202 13:34:40.269816 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:34:40 crc kubenswrapper[4725]: E1202 13:34:40.271881 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:34:48 crc kubenswrapper[4725]: I1202 13:34:48.411193 4725 generic.go:334] "Generic (PLEG): container finished" podID="375b1a65-e012-4cfa-a08b-5d0b321be9d7" containerID="2699d2b16a8b9dd024e73bc6668d4ff63607a8d33aa522ec5fb5bcbbfc1e51b7" exitCode=0 Dec 02 13:34:48 crc kubenswrapper[4725]: I1202 13:34:48.411272 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" event={"ID":"375b1a65-e012-4cfa-a08b-5d0b321be9d7","Type":"ContainerDied","Data":"2699d2b16a8b9dd024e73bc6668d4ff63607a8d33aa522ec5fb5bcbbfc1e51b7"} Dec 02 13:34:48 crc kubenswrapper[4725]: I1202 13:34:48.860783 4725 scope.go:117] "RemoveContainer" containerID="e2dcb35682ba1ded69f2dbde8efad46005747c258d77b01333bdb79e3dbcd36f" Dec 02 13:34:48 crc kubenswrapper[4725]: I1202 13:34:48.883223 4725 scope.go:117] "RemoveContainer" containerID="6e46b5e899fd2cf06fb8a020444ddc795c662bf252af91c3b2cad06068472703" Dec 02 13:34:48 crc kubenswrapper[4725]: I1202 13:34:48.941604 4725 scope.go:117] "RemoveContainer" containerID="0811559d7ba7397b452caaff96197756e30abbfff1f08de1ac4ff06212495fa9" Dec 02 13:34:48 crc kubenswrapper[4725]: I1202 13:34:48.995775 4725 scope.go:117] "RemoveContainer" containerID="942f19c9cb8648563e6cc8eec3980e070e4e79e53f79d2029cd65413b9468f41" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.056489 4725 scope.go:117] "RemoveContainer" containerID="a736568863b41adcae7a1f3fbe454c3b1b0a859bb9d14317860e4451a802abc2" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.079124 4725 scope.go:117] "RemoveContainer" containerID="cae6e01996efca918308c0ea83db409419a9d0c9b4ec7ddadc172962a6884fa7" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.130385 4725 scope.go:117] "RemoveContainer" containerID="a64a2dc372541bc13f6ceef47b73d7c01580d1c10a6a54d32eed66cbe9ff2e49" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.721108 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.860706 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-inventory\") pod \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.860865 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn725\" (UniqueName: \"kubernetes.io/projected/375b1a65-e012-4cfa-a08b-5d0b321be9d7-kube-api-access-pn725\") pod \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.860960 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-ssh-key\") pod \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\" (UID: \"375b1a65-e012-4cfa-a08b-5d0b321be9d7\") " Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.868018 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/375b1a65-e012-4cfa-a08b-5d0b321be9d7-kube-api-access-pn725" (OuterVolumeSpecName: "kube-api-access-pn725") pod "375b1a65-e012-4cfa-a08b-5d0b321be9d7" (UID: "375b1a65-e012-4cfa-a08b-5d0b321be9d7"). InnerVolumeSpecName "kube-api-access-pn725". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.889566 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-inventory" (OuterVolumeSpecName: "inventory") pod "375b1a65-e012-4cfa-a08b-5d0b321be9d7" (UID: "375b1a65-e012-4cfa-a08b-5d0b321be9d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.904667 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "375b1a65-e012-4cfa-a08b-5d0b321be9d7" (UID: "375b1a65-e012-4cfa-a08b-5d0b321be9d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.963504 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn725\" (UniqueName: \"kubernetes.io/projected/375b1a65-e012-4cfa-a08b-5d0b321be9d7-kube-api-access-pn725\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.963532 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:49 crc kubenswrapper[4725]: I1202 13:34:49.963544 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/375b1a65-e012-4cfa-a08b-5d0b321be9d7-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.429530 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" event={"ID":"375b1a65-e012-4cfa-a08b-5d0b321be9d7","Type":"ContainerDied","Data":"27637c00d7bae4fda9edc3433f58505a0316dd40c236d0898f40355193918626"} Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.429884 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27637c00d7bae4fda9edc3433f58505a0316dd40c236d0898f40355193918626" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.429622 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-bxpzs" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.540786 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr"] Dec 02 13:34:50 crc kubenswrapper[4725]: E1202 13:34:50.541387 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="375b1a65-e012-4cfa-a08b-5d0b321be9d7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.541410 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="375b1a65-e012-4cfa-a08b-5d0b321be9d7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.541724 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="375b1a65-e012-4cfa-a08b-5d0b321be9d7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.542650 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.544305 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.545259 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.545435 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.545676 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.550268 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr"] Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.591799 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.591931 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86jnf\" (UniqueName: \"kubernetes.io/projected/b013516f-cd7c-4a91-bc3e-bcddd68f439f-kube-api-access-86jnf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.592249 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.693791 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86jnf\" (UniqueName: \"kubernetes.io/projected/b013516f-cd7c-4a91-bc3e-bcddd68f439f-kube-api-access-86jnf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.693993 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.694063 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.698929 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.701504 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.709892 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86jnf\" (UniqueName: \"kubernetes.io/projected/b013516f-cd7c-4a91-bc3e-bcddd68f439f-kube-api-access-86jnf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-d98hr\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:50 crc kubenswrapper[4725]: I1202 13:34:50.915703 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:34:51 crc kubenswrapper[4725]: I1202 13:34:51.043036 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-cltvz"] Dec 02 13:34:51 crc kubenswrapper[4725]: I1202 13:34:51.054846 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-cltvz"] Dec 02 13:34:51 crc kubenswrapper[4725]: I1202 13:34:51.280226 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f6eb9bb-6946-4475-9421-8ef60d3cb50b" path="/var/lib/kubelet/pods/6f6eb9bb-6946-4475-9421-8ef60d3cb50b/volumes" Dec 02 13:34:51 crc kubenswrapper[4725]: I1202 13:34:51.431354 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr"] Dec 02 13:34:52 crc kubenswrapper[4725]: I1202 13:34:52.032231 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-sffb7"] Dec 02 13:34:52 crc kubenswrapper[4725]: I1202 13:34:52.045002 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-sffb7"] Dec 02 13:34:52 crc kubenswrapper[4725]: I1202 13:34:52.461185 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" event={"ID":"b013516f-cd7c-4a91-bc3e-bcddd68f439f","Type":"ContainerStarted","Data":"2091195490adb63ec1cdfa56f1e487d10e86045b604c1d8e328b44c7e03c7d54"} Dec 02 13:34:52 crc kubenswrapper[4725]: I1202 13:34:52.461491 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" event={"ID":"b013516f-cd7c-4a91-bc3e-bcddd68f439f","Type":"ContainerStarted","Data":"341e6d3b237a3c781e77c4e1036422bdd8281448ce6cbfef9df1aee9f1544176"} Dec 02 13:34:52 crc kubenswrapper[4725]: I1202 13:34:52.481632 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" podStartSLOduration=1.8433386729999999 podStartE2EDuration="2.481612409s" podCreationTimestamp="2025-12-02 13:34:50 +0000 UTC" firstStartedPulling="2025-12-02 13:34:51.439382029 +0000 UTC m=+1822.396023724" lastFinishedPulling="2025-12-02 13:34:52.077655755 +0000 UTC m=+1823.034297460" observedRunningTime="2025-12-02 13:34:52.475877336 +0000 UTC m=+1823.432519031" watchObservedRunningTime="2025-12-02 13:34:52.481612409 +0000 UTC m=+1823.438254104" Dec 02 13:34:53 crc kubenswrapper[4725]: I1202 13:34:53.278710 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a606a54a-2bb1-4390-8682-dd7e78b4e955" path="/var/lib/kubelet/pods/a606a54a-2bb1-4390-8682-dd7e78b4e955/volumes" Dec 02 13:34:55 crc kubenswrapper[4725]: I1202 13:34:55.268176 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:34:56 crc kubenswrapper[4725]: I1202 13:34:56.493971 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"dec08dfad0c2e314494f251f9651e69a5944d3ad6d3e3ff5c1d452de251ae268"} Dec 02 13:35:36 crc kubenswrapper[4725]: I1202 13:35:36.072252 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-92xms"] Dec 02 13:35:36 crc kubenswrapper[4725]: I1202 13:35:36.084651 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-92xms"] Dec 02 13:35:37 crc kubenswrapper[4725]: I1202 13:35:37.286265 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98076d4d-dc13-4a90-9deb-72dc6d9b6095" path="/var/lib/kubelet/pods/98076d4d-dc13-4a90-9deb-72dc6d9b6095/volumes" Dec 02 13:35:37 crc kubenswrapper[4725]: I1202 13:35:37.841200 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" event={"ID":"b013516f-cd7c-4a91-bc3e-bcddd68f439f","Type":"ContainerDied","Data":"2091195490adb63ec1cdfa56f1e487d10e86045b604c1d8e328b44c7e03c7d54"} Dec 02 13:35:37 crc kubenswrapper[4725]: I1202 13:35:37.841064 4725 generic.go:334] "Generic (PLEG): container finished" podID="b013516f-cd7c-4a91-bc3e-bcddd68f439f" containerID="2091195490adb63ec1cdfa56f1e487d10e86045b604c1d8e328b44c7e03c7d54" exitCode=0 Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.275026 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.405053 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-inventory\") pod \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.405249 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-ssh-key\") pod \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.405331 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86jnf\" (UniqueName: \"kubernetes.io/projected/b013516f-cd7c-4a91-bc3e-bcddd68f439f-kube-api-access-86jnf\") pod \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\" (UID: \"b013516f-cd7c-4a91-bc3e-bcddd68f439f\") " Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.412728 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b013516f-cd7c-4a91-bc3e-bcddd68f439f-kube-api-access-86jnf" (OuterVolumeSpecName: "kube-api-access-86jnf") pod "b013516f-cd7c-4a91-bc3e-bcddd68f439f" (UID: "b013516f-cd7c-4a91-bc3e-bcddd68f439f"). InnerVolumeSpecName "kube-api-access-86jnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.433143 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b013516f-cd7c-4a91-bc3e-bcddd68f439f" (UID: "b013516f-cd7c-4a91-bc3e-bcddd68f439f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.433709 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-inventory" (OuterVolumeSpecName: "inventory") pod "b013516f-cd7c-4a91-bc3e-bcddd68f439f" (UID: "b013516f-cd7c-4a91-bc3e-bcddd68f439f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.507590 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.507630 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b013516f-cd7c-4a91-bc3e-bcddd68f439f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.507639 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86jnf\" (UniqueName: \"kubernetes.io/projected/b013516f-cd7c-4a91-bc3e-bcddd68f439f-kube-api-access-86jnf\") on node \"crc\" DevicePath \"\"" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.862108 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" event={"ID":"b013516f-cd7c-4a91-bc3e-bcddd68f439f","Type":"ContainerDied","Data":"341e6d3b237a3c781e77c4e1036422bdd8281448ce6cbfef9df1aee9f1544176"} Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.862148 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="341e6d3b237a3c781e77c4e1036422bdd8281448ce6cbfef9df1aee9f1544176" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.862171 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-d98hr" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.943181 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9shf4"] Dec 02 13:35:39 crc kubenswrapper[4725]: E1202 13:35:39.943668 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b013516f-cd7c-4a91-bc3e-bcddd68f439f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.943691 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b013516f-cd7c-4a91-bc3e-bcddd68f439f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.943896 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b013516f-cd7c-4a91-bc3e-bcddd68f439f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.944538 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.946903 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.946970 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.947186 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.951126 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:35:39 crc kubenswrapper[4725]: I1202 13:35:39.960713 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9shf4"] Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.116844 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrkh2\" (UniqueName: \"kubernetes.io/projected/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-kube-api-access-lrkh2\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.116907 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.116992 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.218660 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrkh2\" (UniqueName: \"kubernetes.io/projected/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-kube-api-access-lrkh2\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.218710 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.218783 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.224547 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.224723 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.238251 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrkh2\" (UniqueName: \"kubernetes.io/projected/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-kube-api-access-lrkh2\") pod \"ssh-known-hosts-edpm-deployment-9shf4\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.264541 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.822611 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9shf4"] Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.823215 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:35:40 crc kubenswrapper[4725]: I1202 13:35:40.872005 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" event={"ID":"50bb29f4-5ab8-421c-a5f5-4321d106e1fe","Type":"ContainerStarted","Data":"11683a57f6552674abd0886ade10f162fdf8329df8073d83016a7e4bc91e5402"} Dec 02 13:35:42 crc kubenswrapper[4725]: I1202 13:35:42.890366 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" event={"ID":"50bb29f4-5ab8-421c-a5f5-4321d106e1fe","Type":"ContainerStarted","Data":"93417ec52b27ed1e3485afb7182f925ae54d8bbd31bab72d8535399ff5f06438"} Dec 02 13:35:49 crc kubenswrapper[4725]: I1202 13:35:49.267799 4725 scope.go:117] "RemoveContainer" containerID="db5cff9dcd14e1d46ebb9ccec66be80fb821bdc98fcb6c2d1debf9cff2c21a17" Dec 02 13:35:49 crc kubenswrapper[4725]: I1202 13:35:49.781975 4725 scope.go:117] "RemoveContainer" containerID="ae0806dcd2c174d69752d508447b56df6d47bbfcb915221751835a7355a1234a" Dec 02 13:35:49 crc kubenswrapper[4725]: I1202 13:35:49.855598 4725 scope.go:117] "RemoveContainer" containerID="2f7f59366c6df9411bd2995749cd6abbc1d98a0170ea697ab1955c05a1b0f28e" Dec 02 13:35:49 crc kubenswrapper[4725]: I1202 13:35:49.947016 4725 generic.go:334] "Generic (PLEG): container finished" podID="50bb29f4-5ab8-421c-a5f5-4321d106e1fe" containerID="93417ec52b27ed1e3485afb7182f925ae54d8bbd31bab72d8535399ff5f06438" exitCode=0 Dec 02 13:35:49 crc kubenswrapper[4725]: I1202 13:35:49.947074 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" event={"ID":"50bb29f4-5ab8-421c-a5f5-4321d106e1fe","Type":"ContainerDied","Data":"93417ec52b27ed1e3485afb7182f925ae54d8bbd31bab72d8535399ff5f06438"} Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.321416 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.464753 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrkh2\" (UniqueName: \"kubernetes.io/projected/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-kube-api-access-lrkh2\") pod \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.464817 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-ssh-key-openstack-edpm-ipam\") pod \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.464916 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-inventory-0\") pod \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\" (UID: \"50bb29f4-5ab8-421c-a5f5-4321d106e1fe\") " Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.470591 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-kube-api-access-lrkh2" (OuterVolumeSpecName: "kube-api-access-lrkh2") pod "50bb29f4-5ab8-421c-a5f5-4321d106e1fe" (UID: "50bb29f4-5ab8-421c-a5f5-4321d106e1fe"). InnerVolumeSpecName "kube-api-access-lrkh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.495508 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "50bb29f4-5ab8-421c-a5f5-4321d106e1fe" (UID: "50bb29f4-5ab8-421c-a5f5-4321d106e1fe"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.499385 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "50bb29f4-5ab8-421c-a5f5-4321d106e1fe" (UID: "50bb29f4-5ab8-421c-a5f5-4321d106e1fe"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.567401 4725 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.567450 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrkh2\" (UniqueName: \"kubernetes.io/projected/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-kube-api-access-lrkh2\") on node \"crc\" DevicePath \"\"" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.567509 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/50bb29f4-5ab8-421c-a5f5-4321d106e1fe-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.963364 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" event={"ID":"50bb29f4-5ab8-421c-a5f5-4321d106e1fe","Type":"ContainerDied","Data":"11683a57f6552674abd0886ade10f162fdf8329df8073d83016a7e4bc91e5402"} Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.963741 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11683a57f6552674abd0886ade10f162fdf8329df8073d83016a7e4bc91e5402" Dec 02 13:35:51 crc kubenswrapper[4725]: I1202 13:35:51.963379 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9shf4" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.030692 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w"] Dec 02 13:35:52 crc kubenswrapper[4725]: E1202 13:35:52.031168 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50bb29f4-5ab8-421c-a5f5-4321d106e1fe" containerName="ssh-known-hosts-edpm-deployment" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.031193 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="50bb29f4-5ab8-421c-a5f5-4321d106e1fe" containerName="ssh-known-hosts-edpm-deployment" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.031483 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="50bb29f4-5ab8-421c-a5f5-4321d106e1fe" containerName="ssh-known-hosts-edpm-deployment" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.032233 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.034048 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.034051 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.034299 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.034405 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.041068 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w"] Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.178040 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.178156 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.178211 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk4xp\" (UniqueName: \"kubernetes.io/projected/2c170bed-2666-437d-b8dc-b14bb81bd6eb-kube-api-access-bk4xp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.279639 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.279774 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.279814 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk4xp\" (UniqueName: \"kubernetes.io/projected/2c170bed-2666-437d-b8dc-b14bb81bd6eb-kube-api-access-bk4xp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.285021 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.285084 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.297371 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk4xp\" (UniqueName: \"kubernetes.io/projected/2c170bed-2666-437d-b8dc-b14bb81bd6eb-kube-api-access-bk4xp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-twz8w\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.353019 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.854334 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w"] Dec 02 13:35:52 crc kubenswrapper[4725]: I1202 13:35:52.973340 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" event={"ID":"2c170bed-2666-437d-b8dc-b14bb81bd6eb","Type":"ContainerStarted","Data":"4d52cc465694f6bd5d06ffeb2a3617ce4ebb1c66c8d8b4aa4699b480fd3c49e7"} Dec 02 13:35:53 crc kubenswrapper[4725]: I1202 13:35:53.982611 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" event={"ID":"2c170bed-2666-437d-b8dc-b14bb81bd6eb","Type":"ContainerStarted","Data":"77e3c9b7c8d418dc8bec1073c3392a58833fdc2432d5e9685f1cacc59b8d831a"} Dec 02 13:35:54 crc kubenswrapper[4725]: I1202 13:35:54.004573 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" podStartSLOduration=1.458892924 podStartE2EDuration="2.004549693s" podCreationTimestamp="2025-12-02 13:35:52 +0000 UTC" firstStartedPulling="2025-12-02 13:35:52.864991688 +0000 UTC m=+1883.821633383" lastFinishedPulling="2025-12-02 13:35:53.410648457 +0000 UTC m=+1884.367290152" observedRunningTime="2025-12-02 13:35:53.998049511 +0000 UTC m=+1884.954691216" watchObservedRunningTime="2025-12-02 13:35:54.004549693 +0000 UTC m=+1884.961191388" Dec 02 13:36:01 crc kubenswrapper[4725]: I1202 13:36:01.041820 4725 generic.go:334] "Generic (PLEG): container finished" podID="2c170bed-2666-437d-b8dc-b14bb81bd6eb" containerID="77e3c9b7c8d418dc8bec1073c3392a58833fdc2432d5e9685f1cacc59b8d831a" exitCode=0 Dec 02 13:36:01 crc kubenswrapper[4725]: I1202 13:36:01.041918 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" event={"ID":"2c170bed-2666-437d-b8dc-b14bb81bd6eb","Type":"ContainerDied","Data":"77e3c9b7c8d418dc8bec1073c3392a58833fdc2432d5e9685f1cacc59b8d831a"} Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.442600 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.486946 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-ssh-key\") pod \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.487135 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk4xp\" (UniqueName: \"kubernetes.io/projected/2c170bed-2666-437d-b8dc-b14bb81bd6eb-kube-api-access-bk4xp\") pod \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.487207 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-inventory\") pod \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\" (UID: \"2c170bed-2666-437d-b8dc-b14bb81bd6eb\") " Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.494113 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c170bed-2666-437d-b8dc-b14bb81bd6eb-kube-api-access-bk4xp" (OuterVolumeSpecName: "kube-api-access-bk4xp") pod "2c170bed-2666-437d-b8dc-b14bb81bd6eb" (UID: "2c170bed-2666-437d-b8dc-b14bb81bd6eb"). InnerVolumeSpecName "kube-api-access-bk4xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.522643 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-inventory" (OuterVolumeSpecName: "inventory") pod "2c170bed-2666-437d-b8dc-b14bb81bd6eb" (UID: "2c170bed-2666-437d-b8dc-b14bb81bd6eb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.524750 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2c170bed-2666-437d-b8dc-b14bb81bd6eb" (UID: "2c170bed-2666-437d-b8dc-b14bb81bd6eb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.590149 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk4xp\" (UniqueName: \"kubernetes.io/projected/2c170bed-2666-437d-b8dc-b14bb81bd6eb-kube-api-access-bk4xp\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.590269 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:02 crc kubenswrapper[4725]: I1202 13:36:02.590348 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2c170bed-2666-437d-b8dc-b14bb81bd6eb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.070667 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" event={"ID":"2c170bed-2666-437d-b8dc-b14bb81bd6eb","Type":"ContainerDied","Data":"4d52cc465694f6bd5d06ffeb2a3617ce4ebb1c66c8d8b4aa4699b480fd3c49e7"} Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.070710 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d52cc465694f6bd5d06ffeb2a3617ce4ebb1c66c8d8b4aa4699b480fd3c49e7" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.070782 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-twz8w" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.160717 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd"] Dec 02 13:36:03 crc kubenswrapper[4725]: E1202 13:36:03.161124 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c170bed-2666-437d-b8dc-b14bb81bd6eb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.161142 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c170bed-2666-437d-b8dc-b14bb81bd6eb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.161344 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c170bed-2666-437d-b8dc-b14bb81bd6eb" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.163511 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.165833 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.166057 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.166283 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.175190 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd"] Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.175986 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.305567 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.305731 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gcx9\" (UniqueName: \"kubernetes.io/projected/0372a2f9-72d1-468a-bd81-5a1a0a48c545-kube-api-access-4gcx9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.305796 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.407921 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.408041 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gcx9\" (UniqueName: \"kubernetes.io/projected/0372a2f9-72d1-468a-bd81-5a1a0a48c545-kube-api-access-4gcx9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.408070 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.413873 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.421597 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.429849 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gcx9\" (UniqueName: \"kubernetes.io/projected/0372a2f9-72d1-468a-bd81-5a1a0a48c545-kube-api-access-4gcx9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.486871 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:03 crc kubenswrapper[4725]: I1202 13:36:03.987313 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd"] Dec 02 13:36:04 crc kubenswrapper[4725]: I1202 13:36:04.080884 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" event={"ID":"0372a2f9-72d1-468a-bd81-5a1a0a48c545","Type":"ContainerStarted","Data":"7774cede9d131faa5540664b866298b66a8f498fd0a6947be0da98a12dab1fab"} Dec 02 13:36:05 crc kubenswrapper[4725]: I1202 13:36:05.090986 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" event={"ID":"0372a2f9-72d1-468a-bd81-5a1a0a48c545","Type":"ContainerStarted","Data":"c3fbca88af1b2d7433ac43cccc7f0933138c233a8856b7dd4d42889416477147"} Dec 02 13:36:05 crc kubenswrapper[4725]: I1202 13:36:05.109375 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" podStartSLOduration=1.293836421 podStartE2EDuration="2.109356271s" podCreationTimestamp="2025-12-02 13:36:03 +0000 UTC" firstStartedPulling="2025-12-02 13:36:03.987827117 +0000 UTC m=+1894.944468812" lastFinishedPulling="2025-12-02 13:36:04.803346977 +0000 UTC m=+1895.759988662" observedRunningTime="2025-12-02 13:36:05.10531545 +0000 UTC m=+1896.061957145" watchObservedRunningTime="2025-12-02 13:36:05.109356271 +0000 UTC m=+1896.065997966" Dec 02 13:36:14 crc kubenswrapper[4725]: I1202 13:36:14.164856 4725 generic.go:334] "Generic (PLEG): container finished" podID="0372a2f9-72d1-468a-bd81-5a1a0a48c545" containerID="c3fbca88af1b2d7433ac43cccc7f0933138c233a8856b7dd4d42889416477147" exitCode=0 Dec 02 13:36:14 crc kubenswrapper[4725]: I1202 13:36:14.164945 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" event={"ID":"0372a2f9-72d1-468a-bd81-5a1a0a48c545","Type":"ContainerDied","Data":"c3fbca88af1b2d7433ac43cccc7f0933138c233a8856b7dd4d42889416477147"} Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.604216 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.752543 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gcx9\" (UniqueName: \"kubernetes.io/projected/0372a2f9-72d1-468a-bd81-5a1a0a48c545-kube-api-access-4gcx9\") pod \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.753072 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory\") pod \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.753174 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-ssh-key\") pod \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.758692 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0372a2f9-72d1-468a-bd81-5a1a0a48c545-kube-api-access-4gcx9" (OuterVolumeSpecName: "kube-api-access-4gcx9") pod "0372a2f9-72d1-468a-bd81-5a1a0a48c545" (UID: "0372a2f9-72d1-468a-bd81-5a1a0a48c545"). InnerVolumeSpecName "kube-api-access-4gcx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:15 crc kubenswrapper[4725]: E1202 13:36:15.778763 4725 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory podName:0372a2f9-72d1-468a-bd81-5a1a0a48c545 nodeName:}" failed. No retries permitted until 2025-12-02 13:36:16.278735387 +0000 UTC m=+1907.235377082 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory") pod "0372a2f9-72d1-468a-bd81-5a1a0a48c545" (UID: "0372a2f9-72d1-468a-bd81-5a1a0a48c545") : error deleting /var/lib/kubelet/pods/0372a2f9-72d1-468a-bd81-5a1a0a48c545/volume-subpaths: remove /var/lib/kubelet/pods/0372a2f9-72d1-468a-bd81-5a1a0a48c545/volume-subpaths: no such file or directory Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.781576 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0372a2f9-72d1-468a-bd81-5a1a0a48c545" (UID: "0372a2f9-72d1-468a-bd81-5a1a0a48c545"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.855617 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gcx9\" (UniqueName: \"kubernetes.io/projected/0372a2f9-72d1-468a-bd81-5a1a0a48c545-kube-api-access-4gcx9\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:15 crc kubenswrapper[4725]: I1202 13:36:15.855679 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.181496 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" event={"ID":"0372a2f9-72d1-468a-bd81-5a1a0a48c545","Type":"ContainerDied","Data":"7774cede9d131faa5540664b866298b66a8f498fd0a6947be0da98a12dab1fab"} Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.181542 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7774cede9d131faa5540664b866298b66a8f498fd0a6947be0da98a12dab1fab" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.181620 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.263021 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8"] Dec 02 13:36:16 crc kubenswrapper[4725]: E1202 13:36:16.263480 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372a2f9-72d1-468a-bd81-5a1a0a48c545" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.263503 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372a2f9-72d1-468a-bd81-5a1a0a48c545" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.263787 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="0372a2f9-72d1-468a-bd81-5a1a0a48c545" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.264595 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.267054 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.267506 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.267824 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.268635 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.280123 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8"] Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.363905 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory\") pod \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\" (UID: \"0372a2f9-72d1-468a-bd81-5a1a0a48c545\") " Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.364573 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5dxk\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-kube-api-access-g5dxk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.364695 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.364747 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.364867 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.364949 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365083 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365175 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365445 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365498 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365661 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365702 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365739 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365774 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.365807 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.373663 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory" (OuterVolumeSpecName: "inventory") pod "0372a2f9-72d1-468a-bd81-5a1a0a48c545" (UID: "0372a2f9-72d1-468a-bd81-5a1a0a48c545"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.467479 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.467791 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.467980 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5dxk\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-kube-api-access-g5dxk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468120 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468245 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468387 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468492 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468623 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468746 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468883 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.468996 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.469097 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.469203 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.469305 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.469434 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0372a2f9-72d1-468a-bd81-5a1a0a48c545-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.472829 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.473700 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.473726 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.474241 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.474394 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.475814 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.476089 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.476183 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.476184 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.476293 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.477310 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.481337 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.484281 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.488600 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5dxk\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-kube-api-access-g5dxk\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:16 crc kubenswrapper[4725]: I1202 13:36:16.582964 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:17 crc kubenswrapper[4725]: I1202 13:36:17.088498 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8"] Dec 02 13:36:17 crc kubenswrapper[4725]: I1202 13:36:17.193838 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" event={"ID":"5fee6c15-b89f-4d6b-8536-0e14c298e235","Type":"ContainerStarted","Data":"6d396dcb962d31653c628075a4be99043e8ae72ed1089ea0d136c6d87a11292e"} Dec 02 13:36:18 crc kubenswrapper[4725]: I1202 13:36:18.202834 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" event={"ID":"5fee6c15-b89f-4d6b-8536-0e14c298e235","Type":"ContainerStarted","Data":"4dc9f1d215c8096f49d05b2c583ba7016c7c4b4f0f3fc13bd07c83e6abc349e4"} Dec 02 13:36:18 crc kubenswrapper[4725]: I1202 13:36:18.230433 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" podStartSLOduration=1.493733527 podStartE2EDuration="2.230411634s" podCreationTimestamp="2025-12-02 13:36:16 +0000 UTC" firstStartedPulling="2025-12-02 13:36:17.093630308 +0000 UTC m=+1908.050272003" lastFinishedPulling="2025-12-02 13:36:17.830308415 +0000 UTC m=+1908.786950110" observedRunningTime="2025-12-02 13:36:18.22308846 +0000 UTC m=+1909.179730155" watchObservedRunningTime="2025-12-02 13:36:18.230411634 +0000 UTC m=+1909.187053329" Dec 02 13:36:54 crc kubenswrapper[4725]: I1202 13:36:54.513300 4725 generic.go:334] "Generic (PLEG): container finished" podID="5fee6c15-b89f-4d6b-8536-0e14c298e235" containerID="4dc9f1d215c8096f49d05b2c583ba7016c7c4b4f0f3fc13bd07c83e6abc349e4" exitCode=0 Dec 02 13:36:54 crc kubenswrapper[4725]: I1202 13:36:54.513376 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" event={"ID":"5fee6c15-b89f-4d6b-8536-0e14c298e235","Type":"ContainerDied","Data":"4dc9f1d215c8096f49d05b2c583ba7016c7c4b4f0f3fc13bd07c83e6abc349e4"} Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.084759 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226321 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-telemetry-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226418 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-repo-setup-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226448 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ssh-key\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226563 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ovn-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226617 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-inventory\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226641 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-nova-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226670 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-neutron-metadata-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226724 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.226764 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5dxk\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-kube-api-access-g5dxk\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.227433 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-ovn-default-certs-0\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.227497 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.227555 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-libvirt-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.227616 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.227650 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-bootstrap-combined-ca-bundle\") pod \"5fee6c15-b89f-4d6b-8536-0e14c298e235\" (UID: \"5fee6c15-b89f-4d6b-8536-0e14c298e235\") " Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.234338 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.235376 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.235424 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.235481 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.236545 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.236612 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.237617 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.237829 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.238204 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-kube-api-access-g5dxk" (OuterVolumeSpecName: "kube-api-access-g5dxk") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "kube-api-access-g5dxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.238682 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.239231 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.247192 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.262367 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-inventory" (OuterVolumeSpecName: "inventory") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.262543 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fee6c15-b89f-4d6b-8536-0e14c298e235" (UID: "5fee6c15-b89f-4d6b-8536-0e14c298e235"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330157 4725 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330188 4725 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330199 4725 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330208 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330220 4725 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330230 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330239 4725 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330247 4725 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330255 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330265 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5dxk\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-kube-api-access-g5dxk\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330276 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330287 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330297 4725 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fee6c15-b89f-4d6b-8536-0e14c298e235-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.330308 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5fee6c15-b89f-4d6b-8536-0e14c298e235-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.533070 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" event={"ID":"5fee6c15-b89f-4d6b-8536-0e14c298e235","Type":"ContainerDied","Data":"6d396dcb962d31653c628075a4be99043e8ae72ed1089ea0d136c6d87a11292e"} Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.533117 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d396dcb962d31653c628075a4be99043e8ae72ed1089ea0d136c6d87a11292e" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.533427 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.634375 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6"] Dec 02 13:36:56 crc kubenswrapper[4725]: E1202 13:36:56.634914 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fee6c15-b89f-4d6b-8536-0e14c298e235" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.634940 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fee6c15-b89f-4d6b-8536-0e14c298e235" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.635175 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fee6c15-b89f-4d6b-8536-0e14c298e235" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.636034 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.638663 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.639040 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.639196 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.639251 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.641079 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.652544 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6"] Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.740002 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.740101 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.740265 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.740300 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2gw4\" (UniqueName: \"kubernetes.io/projected/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-kube-api-access-s2gw4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.740359 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.842375 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.842438 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2gw4\" (UniqueName: \"kubernetes.io/projected/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-kube-api-access-s2gw4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.842586 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.842673 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.842732 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.844070 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.846996 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.847012 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.848069 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.859920 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2gw4\" (UniqueName: \"kubernetes.io/projected/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-kube-api-access-s2gw4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-rlrw6\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:56 crc kubenswrapper[4725]: I1202 13:36:56.972858 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:36:57 crc kubenswrapper[4725]: W1202 13:36:57.457055 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9deb432a_a6c8_4a8d_9b5e_bf50fa45cee1.slice/crio-b5db3f2ed09fe70556f03bf3daa0f09e53fc33e233c5be78ac6e86d31006d9ff WatchSource:0}: Error finding container b5db3f2ed09fe70556f03bf3daa0f09e53fc33e233c5be78ac6e86d31006d9ff: Status 404 returned error can't find the container with id b5db3f2ed09fe70556f03bf3daa0f09e53fc33e233c5be78ac6e86d31006d9ff Dec 02 13:36:57 crc kubenswrapper[4725]: I1202 13:36:57.458307 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6"] Dec 02 13:36:57 crc kubenswrapper[4725]: I1202 13:36:57.542901 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" event={"ID":"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1","Type":"ContainerStarted","Data":"b5db3f2ed09fe70556f03bf3daa0f09e53fc33e233c5be78ac6e86d31006d9ff"} Dec 02 13:36:58 crc kubenswrapper[4725]: I1202 13:36:58.552540 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" event={"ID":"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1","Type":"ContainerStarted","Data":"67e4df175eceb63a188c62a5b45de4e74bb97a7674c2537787e24e9df60b88a8"} Dec 02 13:36:58 crc kubenswrapper[4725]: I1202 13:36:58.578362 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" podStartSLOduration=2.009079972 podStartE2EDuration="2.578339992s" podCreationTimestamp="2025-12-02 13:36:56 +0000 UTC" firstStartedPulling="2025-12-02 13:36:57.459618017 +0000 UTC m=+1948.416259712" lastFinishedPulling="2025-12-02 13:36:58.028878037 +0000 UTC m=+1948.985519732" observedRunningTime="2025-12-02 13:36:58.567778567 +0000 UTC m=+1949.524420262" watchObservedRunningTime="2025-12-02 13:36:58.578339992 +0000 UTC m=+1949.534981687" Dec 02 13:37:24 crc kubenswrapper[4725]: I1202 13:37:24.285261 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:37:24 crc kubenswrapper[4725]: I1202 13:37:24.286343 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.593241 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9nbd5"] Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.595844 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.604162 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9nbd5"] Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.785997 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-utilities\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.786087 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-catalog-content\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.786607 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x99zk\" (UniqueName: \"kubernetes.io/projected/89d65328-4a50-4ff2-b909-10e633d2bb78-kube-api-access-x99zk\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.888263 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-utilities\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.888682 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-catalog-content\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.888908 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x99zk\" (UniqueName: \"kubernetes.io/projected/89d65328-4a50-4ff2-b909-10e633d2bb78-kube-api-access-x99zk\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.889028 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-catalog-content\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.888931 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-utilities\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.911644 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x99zk\" (UniqueName: \"kubernetes.io/projected/89d65328-4a50-4ff2-b909-10e633d2bb78-kube-api-access-x99zk\") pod \"redhat-marketplace-9nbd5\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:41 crc kubenswrapper[4725]: I1202 13:37:41.917055 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:42 crc kubenswrapper[4725]: I1202 13:37:42.357853 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9nbd5"] Dec 02 13:37:43 crc kubenswrapper[4725]: I1202 13:37:43.020062 4725 generic.go:334] "Generic (PLEG): container finished" podID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerID="c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed" exitCode=0 Dec 02 13:37:43 crc kubenswrapper[4725]: I1202 13:37:43.020250 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerDied","Data":"c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed"} Dec 02 13:37:43 crc kubenswrapper[4725]: I1202 13:37:43.020389 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerStarted","Data":"b352b65bb1319d2ddb0da3baf5fbe651d1629bd8e2947f11ba1d86aaff3d1102"} Dec 02 13:37:44 crc kubenswrapper[4725]: I1202 13:37:44.031156 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerStarted","Data":"6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74"} Dec 02 13:37:45 crc kubenswrapper[4725]: I1202 13:37:45.045253 4725 generic.go:334] "Generic (PLEG): container finished" podID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerID="6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74" exitCode=0 Dec 02 13:37:45 crc kubenswrapper[4725]: I1202 13:37:45.045732 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerDied","Data":"6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74"} Dec 02 13:37:46 crc kubenswrapper[4725]: I1202 13:37:46.055399 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerStarted","Data":"14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef"} Dec 02 13:37:46 crc kubenswrapper[4725]: I1202 13:37:46.082415 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9nbd5" podStartSLOduration=2.559807794 podStartE2EDuration="5.082363374s" podCreationTimestamp="2025-12-02 13:37:41 +0000 UTC" firstStartedPulling="2025-12-02 13:37:43.021359595 +0000 UTC m=+1993.978001290" lastFinishedPulling="2025-12-02 13:37:45.543915175 +0000 UTC m=+1996.500556870" observedRunningTime="2025-12-02 13:37:46.07419603 +0000 UTC m=+1997.030837735" watchObservedRunningTime="2025-12-02 13:37:46.082363374 +0000 UTC m=+1997.039005069" Dec 02 13:37:51 crc kubenswrapper[4725]: I1202 13:37:51.917579 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:51 crc kubenswrapper[4725]: I1202 13:37:51.918145 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:51 crc kubenswrapper[4725]: I1202 13:37:51.964138 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:52 crc kubenswrapper[4725]: I1202 13:37:52.150957 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:52 crc kubenswrapper[4725]: I1202 13:37:52.197483 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9nbd5"] Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.125274 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9nbd5" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="registry-server" containerID="cri-o://14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef" gracePeriod=2 Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.285039 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.285354 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.533526 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.640396 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-catalog-content\") pod \"89d65328-4a50-4ff2-b909-10e633d2bb78\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.640481 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x99zk\" (UniqueName: \"kubernetes.io/projected/89d65328-4a50-4ff2-b909-10e633d2bb78-kube-api-access-x99zk\") pod \"89d65328-4a50-4ff2-b909-10e633d2bb78\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.640559 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-utilities\") pod \"89d65328-4a50-4ff2-b909-10e633d2bb78\" (UID: \"89d65328-4a50-4ff2-b909-10e633d2bb78\") " Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.641572 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-utilities" (OuterVolumeSpecName: "utilities") pod "89d65328-4a50-4ff2-b909-10e633d2bb78" (UID: "89d65328-4a50-4ff2-b909-10e633d2bb78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.646257 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89d65328-4a50-4ff2-b909-10e633d2bb78-kube-api-access-x99zk" (OuterVolumeSpecName: "kube-api-access-x99zk") pod "89d65328-4a50-4ff2-b909-10e633d2bb78" (UID: "89d65328-4a50-4ff2-b909-10e633d2bb78"). InnerVolumeSpecName "kube-api-access-x99zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.659726 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89d65328-4a50-4ff2-b909-10e633d2bb78" (UID: "89d65328-4a50-4ff2-b909-10e633d2bb78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.743134 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.743427 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x99zk\" (UniqueName: \"kubernetes.io/projected/89d65328-4a50-4ff2-b909-10e633d2bb78-kube-api-access-x99zk\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:54 crc kubenswrapper[4725]: I1202 13:37:54.743439 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89d65328-4a50-4ff2-b909-10e633d2bb78-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.135000 4725 generic.go:334] "Generic (PLEG): container finished" podID="9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" containerID="67e4df175eceb63a188c62a5b45de4e74bb97a7674c2537787e24e9df60b88a8" exitCode=0 Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.135062 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" event={"ID":"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1","Type":"ContainerDied","Data":"67e4df175eceb63a188c62a5b45de4e74bb97a7674c2537787e24e9df60b88a8"} Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.142716 4725 generic.go:334] "Generic (PLEG): container finished" podID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerID="14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef" exitCode=0 Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.142775 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerDied","Data":"14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef"} Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.142810 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9nbd5" event={"ID":"89d65328-4a50-4ff2-b909-10e633d2bb78","Type":"ContainerDied","Data":"b352b65bb1319d2ddb0da3baf5fbe651d1629bd8e2947f11ba1d86aaff3d1102"} Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.142820 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9nbd5" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.142844 4725 scope.go:117] "RemoveContainer" containerID="14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.177821 4725 scope.go:117] "RemoveContainer" containerID="6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.183293 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9nbd5"] Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.191822 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9nbd5"] Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.201023 4725 scope.go:117] "RemoveContainer" containerID="c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.243748 4725 scope.go:117] "RemoveContainer" containerID="14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef" Dec 02 13:37:55 crc kubenswrapper[4725]: E1202 13:37:55.245025 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef\": container with ID starting with 14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef not found: ID does not exist" containerID="14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.245070 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef"} err="failed to get container status \"14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef\": rpc error: code = NotFound desc = could not find container \"14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef\": container with ID starting with 14d94f52cecbe42565e2b6839e0788f685342510af3e44ae44589f5018393aef not found: ID does not exist" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.245103 4725 scope.go:117] "RemoveContainer" containerID="6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74" Dec 02 13:37:55 crc kubenswrapper[4725]: E1202 13:37:55.245512 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74\": container with ID starting with 6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74 not found: ID does not exist" containerID="6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.245566 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74"} err="failed to get container status \"6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74\": rpc error: code = NotFound desc = could not find container \"6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74\": container with ID starting with 6827913d461ca8dc82f00c29cbad6730a44c57540f3fd2153ba50a92c07b2b74 not found: ID does not exist" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.245598 4725 scope.go:117] "RemoveContainer" containerID="c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed" Dec 02 13:37:55 crc kubenswrapper[4725]: E1202 13:37:55.245955 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed\": container with ID starting with c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed not found: ID does not exist" containerID="c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.245996 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed"} err="failed to get container status \"c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed\": rpc error: code = NotFound desc = could not find container \"c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed\": container with ID starting with c86e9c0b07a17774aa0978b5a360ee94d5a473898dbfedc8d9777934ff984aed not found: ID does not exist" Dec 02 13:37:55 crc kubenswrapper[4725]: I1202 13:37:55.279543 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" path="/var/lib/kubelet/pods/89d65328-4a50-4ff2-b909-10e633d2bb78/volumes" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.564916 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.579402 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ssh-key\") pod \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.579529 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2gw4\" (UniqueName: \"kubernetes.io/projected/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-kube-api-access-s2gw4\") pod \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.579588 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovncontroller-config-0\") pod \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.584320 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-inventory\") pod \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.584409 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovn-combined-ca-bundle\") pod \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\" (UID: \"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1\") " Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.661435 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" (UID: "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.661947 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-kube-api-access-s2gw4" (OuterVolumeSpecName: "kube-api-access-s2gw4") pod "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" (UID: "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1"). InnerVolumeSpecName "kube-api-access-s2gw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.665014 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" (UID: "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.668326 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" (UID: "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.668804 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-inventory" (OuterVolumeSpecName: "inventory") pod "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" (UID: "9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.687695 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.687729 4725 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.687740 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.687752 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2gw4\" (UniqueName: \"kubernetes.io/projected/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-kube-api-access-s2gw4\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:56 crc kubenswrapper[4725]: I1202 13:37:56.687762 4725 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.173178 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" event={"ID":"9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1","Type":"ContainerDied","Data":"b5db3f2ed09fe70556f03bf3daa0f09e53fc33e233c5be78ac6e86d31006d9ff"} Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.173224 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5db3f2ed09fe70556f03bf3daa0f09e53fc33e233c5be78ac6e86d31006d9ff" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.173320 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-rlrw6" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.258497 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk"] Dec 02 13:37:57 crc kubenswrapper[4725]: E1202 13:37:57.259023 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="extract-content" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.259046 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="extract-content" Dec 02 13:37:57 crc kubenswrapper[4725]: E1202 13:37:57.259067 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="extract-utilities" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.259076 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="extract-utilities" Dec 02 13:37:57 crc kubenswrapper[4725]: E1202 13:37:57.259092 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="registry-server" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.259100 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="registry-server" Dec 02 13:37:57 crc kubenswrapper[4725]: E1202 13:37:57.259132 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.259140 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.259368 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.259404 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="89d65328-4a50-4ff2-b909-10e633d2bb78" containerName="registry-server" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.260235 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.267796 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.267953 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.268159 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.268390 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.269047 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.269224 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.286546 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk"] Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.299768 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.299887 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.299934 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.299969 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfvtz\" (UniqueName: \"kubernetes.io/projected/d9633180-0a23-47b6-994f-13ec1eed9c30-kube-api-access-wfvtz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.300130 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.300193 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.402651 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.403071 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.403248 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.403515 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.404371 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.404709 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfvtz\" (UniqueName: \"kubernetes.io/projected/d9633180-0a23-47b6-994f-13ec1eed9c30-kube-api-access-wfvtz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.409199 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.410896 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.411252 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.411564 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.412191 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.422084 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfvtz\" (UniqueName: \"kubernetes.io/projected/d9633180-0a23-47b6-994f-13ec1eed9c30-kube-api-access-wfvtz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.588482 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:37:57 crc kubenswrapper[4725]: I1202 13:37:57.903089 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk"] Dec 02 13:37:58 crc kubenswrapper[4725]: I1202 13:37:58.181808 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" event={"ID":"d9633180-0a23-47b6-994f-13ec1eed9c30","Type":"ContainerStarted","Data":"3309ffd4c8d97db94dd6b66469e2401068d3c3c9cef4c7a580e20509ea54068b"} Dec 02 13:37:59 crc kubenswrapper[4725]: I1202 13:37:59.190157 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" event={"ID":"d9633180-0a23-47b6-994f-13ec1eed9c30","Type":"ContainerStarted","Data":"a8fd5532b82006d95ce1097cfbef0e39a510b5a2c1643d4dc5ac3fa62326d559"} Dec 02 13:37:59 crc kubenswrapper[4725]: I1202 13:37:59.205236 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" podStartSLOduration=1.471467444 podStartE2EDuration="2.205216497s" podCreationTimestamp="2025-12-02 13:37:57 +0000 UTC" firstStartedPulling="2025-12-02 13:37:57.906857152 +0000 UTC m=+2008.863498847" lastFinishedPulling="2025-12-02 13:37:58.640606165 +0000 UTC m=+2009.597247900" observedRunningTime="2025-12-02 13:37:59.204687263 +0000 UTC m=+2010.161328978" watchObservedRunningTime="2025-12-02 13:37:59.205216497 +0000 UTC m=+2010.161858192" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.519997 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hm96j"] Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.522718 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.535349 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hm96j"] Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.648715 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-catalog-content\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.648859 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5ftk\" (UniqueName: \"kubernetes.io/projected/93904807-1cd5-40f8-8111-a063f2bca6a5-kube-api-access-m5ftk\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.649436 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-utilities\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.751083 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-catalog-content\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.751272 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5ftk\" (UniqueName: \"kubernetes.io/projected/93904807-1cd5-40f8-8111-a063f2bca6a5-kube-api-access-m5ftk\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.751314 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-utilities\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.751792 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-utilities\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.751800 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-catalog-content\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.772546 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5ftk\" (UniqueName: \"kubernetes.io/projected/93904807-1cd5-40f8-8111-a063f2bca6a5-kube-api-access-m5ftk\") pod \"redhat-operators-hm96j\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:04 crc kubenswrapper[4725]: I1202 13:38:04.845908 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:05 crc kubenswrapper[4725]: I1202 13:38:05.323681 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hm96j"] Dec 02 13:38:06 crc kubenswrapper[4725]: I1202 13:38:06.244638 4725 generic.go:334] "Generic (PLEG): container finished" podID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerID="edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0" exitCode=0 Dec 02 13:38:06 crc kubenswrapper[4725]: I1202 13:38:06.244763 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerDied","Data":"edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0"} Dec 02 13:38:06 crc kubenswrapper[4725]: I1202 13:38:06.245011 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerStarted","Data":"7926ae39f6965f3b0fa8360d5cdba315d7da69870292256beaff6639f1f2c558"} Dec 02 13:38:08 crc kubenswrapper[4725]: I1202 13:38:08.261487 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerStarted","Data":"ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c"} Dec 02 13:38:11 crc kubenswrapper[4725]: I1202 13:38:11.292649 4725 generic.go:334] "Generic (PLEG): container finished" podID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerID="ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c" exitCode=0 Dec 02 13:38:11 crc kubenswrapper[4725]: I1202 13:38:11.292818 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerDied","Data":"ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c"} Dec 02 13:38:13 crc kubenswrapper[4725]: I1202 13:38:13.312999 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerStarted","Data":"35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594"} Dec 02 13:38:13 crc kubenswrapper[4725]: I1202 13:38:13.330088 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hm96j" podStartSLOduration=2.631345277 podStartE2EDuration="9.330067928s" podCreationTimestamp="2025-12-02 13:38:04 +0000 UTC" firstStartedPulling="2025-12-02 13:38:06.247606323 +0000 UTC m=+2017.204248018" lastFinishedPulling="2025-12-02 13:38:12.946328974 +0000 UTC m=+2023.902970669" observedRunningTime="2025-12-02 13:38:13.329827722 +0000 UTC m=+2024.286469417" watchObservedRunningTime="2025-12-02 13:38:13.330067928 +0000 UTC m=+2024.286709623" Dec 02 13:38:14 crc kubenswrapper[4725]: I1202 13:38:14.846414 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:14 crc kubenswrapper[4725]: I1202 13:38:14.846763 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:15 crc kubenswrapper[4725]: I1202 13:38:15.892084 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hm96j" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="registry-server" probeResult="failure" output=< Dec 02 13:38:15 crc kubenswrapper[4725]: timeout: failed to connect service ":50051" within 1s Dec 02 13:38:15 crc kubenswrapper[4725]: > Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.284935 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.285437 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.285496 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.286346 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dec08dfad0c2e314494f251f9651e69a5944d3ad6d3e3ff5c1d452de251ae268"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.286402 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://dec08dfad0c2e314494f251f9651e69a5944d3ad6d3e3ff5c1d452de251ae268" gracePeriod=600 Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.423438 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="dec08dfad0c2e314494f251f9651e69a5944d3ad6d3e3ff5c1d452de251ae268" exitCode=0 Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.423506 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"dec08dfad0c2e314494f251f9651e69a5944d3ad6d3e3ff5c1d452de251ae268"} Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.423547 4725 scope.go:117] "RemoveContainer" containerID="1ea5c9e6ef8c877889bdf2a7bfdbd3d2560e3fa328084a3c8ddf0dab832aa799" Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.904236 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:24 crc kubenswrapper[4725]: I1202 13:38:24.964576 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:25 crc kubenswrapper[4725]: I1202 13:38:25.141742 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hm96j"] Dec 02 13:38:25 crc kubenswrapper[4725]: I1202 13:38:25.433997 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6"} Dec 02 13:38:26 crc kubenswrapper[4725]: I1202 13:38:26.440981 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hm96j" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="registry-server" containerID="cri-o://35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594" gracePeriod=2 Dec 02 13:38:26 crc kubenswrapper[4725]: I1202 13:38:26.916384 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.040304 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-utilities\") pod \"93904807-1cd5-40f8-8111-a063f2bca6a5\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.040415 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-catalog-content\") pod \"93904807-1cd5-40f8-8111-a063f2bca6a5\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.040604 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5ftk\" (UniqueName: \"kubernetes.io/projected/93904807-1cd5-40f8-8111-a063f2bca6a5-kube-api-access-m5ftk\") pod \"93904807-1cd5-40f8-8111-a063f2bca6a5\" (UID: \"93904807-1cd5-40f8-8111-a063f2bca6a5\") " Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.041256 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-utilities" (OuterVolumeSpecName: "utilities") pod "93904807-1cd5-40f8-8111-a063f2bca6a5" (UID: "93904807-1cd5-40f8-8111-a063f2bca6a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.046411 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93904807-1cd5-40f8-8111-a063f2bca6a5-kube-api-access-m5ftk" (OuterVolumeSpecName: "kube-api-access-m5ftk") pod "93904807-1cd5-40f8-8111-a063f2bca6a5" (UID: "93904807-1cd5-40f8-8111-a063f2bca6a5"). InnerVolumeSpecName "kube-api-access-m5ftk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.143329 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5ftk\" (UniqueName: \"kubernetes.io/projected/93904807-1cd5-40f8-8111-a063f2bca6a5-kube-api-access-m5ftk\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.143665 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.158995 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93904807-1cd5-40f8-8111-a063f2bca6a5" (UID: "93904807-1cd5-40f8-8111-a063f2bca6a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.245516 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93904807-1cd5-40f8-8111-a063f2bca6a5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.452878 4725 generic.go:334] "Generic (PLEG): container finished" podID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerID="35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594" exitCode=0 Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.452925 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerDied","Data":"35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594"} Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.452956 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hm96j" event={"ID":"93904807-1cd5-40f8-8111-a063f2bca6a5","Type":"ContainerDied","Data":"7926ae39f6965f3b0fa8360d5cdba315d7da69870292256beaff6639f1f2c558"} Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.452996 4725 scope.go:117] "RemoveContainer" containerID="35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.454105 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hm96j" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.477041 4725 scope.go:117] "RemoveContainer" containerID="ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.485590 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hm96j"] Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.499067 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hm96j"] Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.504154 4725 scope.go:117] "RemoveContainer" containerID="edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.545758 4725 scope.go:117] "RemoveContainer" containerID="35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594" Dec 02 13:38:27 crc kubenswrapper[4725]: E1202 13:38:27.546284 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594\": container with ID starting with 35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594 not found: ID does not exist" containerID="35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.546385 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594"} err="failed to get container status \"35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594\": rpc error: code = NotFound desc = could not find container \"35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594\": container with ID starting with 35ac9525b489f8f1e9e688e01bc9b3391f15ef4c5745eb9937887e73b9a5f594 not found: ID does not exist" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.546504 4725 scope.go:117] "RemoveContainer" containerID="ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c" Dec 02 13:38:27 crc kubenswrapper[4725]: E1202 13:38:27.546982 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c\": container with ID starting with ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c not found: ID does not exist" containerID="ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.547016 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c"} err="failed to get container status \"ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c\": rpc error: code = NotFound desc = could not find container \"ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c\": container with ID starting with ff468cec0efd472e6801fd3696e1bc39c66489e12213f7c2c850188b91a68f5c not found: ID does not exist" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.547039 4725 scope.go:117] "RemoveContainer" containerID="edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0" Dec 02 13:38:27 crc kubenswrapper[4725]: E1202 13:38:27.547367 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0\": container with ID starting with edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0 not found: ID does not exist" containerID="edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0" Dec 02 13:38:27 crc kubenswrapper[4725]: I1202 13:38:27.547485 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0"} err="failed to get container status \"edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0\": rpc error: code = NotFound desc = could not find container \"edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0\": container with ID starting with edad0ec5fff62f033385d6669035b05ba8a9e3d3760534258bb385491430e4e0 not found: ID does not exist" Dec 02 13:38:29 crc kubenswrapper[4725]: I1202 13:38:29.284800 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" path="/var/lib/kubelet/pods/93904807-1cd5-40f8-8111-a063f2bca6a5/volumes" Dec 02 13:38:42 crc kubenswrapper[4725]: I1202 13:38:42.585498 4725 generic.go:334] "Generic (PLEG): container finished" podID="d9633180-0a23-47b6-994f-13ec1eed9c30" containerID="a8fd5532b82006d95ce1097cfbef0e39a510b5a2c1643d4dc5ac3fa62326d559" exitCode=0 Dec 02 13:38:42 crc kubenswrapper[4725]: I1202 13:38:42.585591 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" event={"ID":"d9633180-0a23-47b6-994f-13ec1eed9c30","Type":"ContainerDied","Data":"a8fd5532b82006d95ce1097cfbef0e39a510b5a2c1643d4dc5ac3fa62326d559"} Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.128256 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.206305 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-ssh-key\") pod \"d9633180-0a23-47b6-994f-13ec1eed9c30\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.207073 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-inventory\") pod \"d9633180-0a23-47b6-994f-13ec1eed9c30\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.207222 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-ovn-metadata-agent-neutron-config-0\") pod \"d9633180-0a23-47b6-994f-13ec1eed9c30\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.207327 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-nova-metadata-neutron-config-0\") pod \"d9633180-0a23-47b6-994f-13ec1eed9c30\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.207423 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfvtz\" (UniqueName: \"kubernetes.io/projected/d9633180-0a23-47b6-994f-13ec1eed9c30-kube-api-access-wfvtz\") pod \"d9633180-0a23-47b6-994f-13ec1eed9c30\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.207825 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-metadata-combined-ca-bundle\") pod \"d9633180-0a23-47b6-994f-13ec1eed9c30\" (UID: \"d9633180-0a23-47b6-994f-13ec1eed9c30\") " Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.213102 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d9633180-0a23-47b6-994f-13ec1eed9c30" (UID: "d9633180-0a23-47b6-994f-13ec1eed9c30"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.217307 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9633180-0a23-47b6-994f-13ec1eed9c30-kube-api-access-wfvtz" (OuterVolumeSpecName: "kube-api-access-wfvtz") pod "d9633180-0a23-47b6-994f-13ec1eed9c30" (UID: "d9633180-0a23-47b6-994f-13ec1eed9c30"). InnerVolumeSpecName "kube-api-access-wfvtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.237034 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-inventory" (OuterVolumeSpecName: "inventory") pod "d9633180-0a23-47b6-994f-13ec1eed9c30" (UID: "d9633180-0a23-47b6-994f-13ec1eed9c30"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.239445 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "d9633180-0a23-47b6-994f-13ec1eed9c30" (UID: "d9633180-0a23-47b6-994f-13ec1eed9c30"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.244013 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d9633180-0a23-47b6-994f-13ec1eed9c30" (UID: "d9633180-0a23-47b6-994f-13ec1eed9c30"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.245005 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "d9633180-0a23-47b6-994f-13ec1eed9c30" (UID: "d9633180-0a23-47b6-994f-13ec1eed9c30"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.310545 4725 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.310582 4725 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.310592 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfvtz\" (UniqueName: \"kubernetes.io/projected/d9633180-0a23-47b6-994f-13ec1eed9c30-kube-api-access-wfvtz\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.310605 4725 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.310614 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.310623 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9633180-0a23-47b6-994f-13ec1eed9c30-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.604598 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" event={"ID":"d9633180-0a23-47b6-994f-13ec1eed9c30","Type":"ContainerDied","Data":"3309ffd4c8d97db94dd6b66469e2401068d3c3c9cef4c7a580e20509ea54068b"} Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.604639 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3309ffd4c8d97db94dd6b66469e2401068d3c3c9cef4c7a580e20509ea54068b" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.604655 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.850467 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx"] Dec 02 13:38:44 crc kubenswrapper[4725]: E1202 13:38:44.851103 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="extract-utilities" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.851126 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="extract-utilities" Dec 02 13:38:44 crc kubenswrapper[4725]: E1202 13:38:44.851146 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="extract-content" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.851154 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="extract-content" Dec 02 13:38:44 crc kubenswrapper[4725]: E1202 13:38:44.851170 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="registry-server" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.851177 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="registry-server" Dec 02 13:38:44 crc kubenswrapper[4725]: E1202 13:38:44.851188 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9633180-0a23-47b6-994f-13ec1eed9c30" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.851201 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9633180-0a23-47b6-994f-13ec1eed9c30" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.851430 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="93904807-1cd5-40f8-8111-a063f2bca6a5" containerName="registry-server" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.851444 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9633180-0a23-47b6-994f-13ec1eed9c30" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.852349 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.855295 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.856056 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.856136 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.857151 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.857973 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:38:44 crc kubenswrapper[4725]: I1202 13:38:44.865395 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx"] Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.022205 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.022264 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qjdv\" (UniqueName: \"kubernetes.io/projected/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-kube-api-access-9qjdv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.022307 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.022358 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.022398 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.125505 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.125706 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.125746 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qjdv\" (UniqueName: \"kubernetes.io/projected/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-kube-api-access-9qjdv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.125795 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.125844 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.132212 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.133126 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.141536 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.142930 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.151632 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qjdv\" (UniqueName: \"kubernetes.io/projected/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-kube-api-access-9qjdv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.237102 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:38:45 crc kubenswrapper[4725]: W1202 13:38:45.906367 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee65f1b9_bbb7_4561_af64_e6f0ab3deec1.slice/crio-4440ed44f6d92136c8d6dff7a26889f8dc9461aecfa7b7ba13cef0039df3df13 WatchSource:0}: Error finding container 4440ed44f6d92136c8d6dff7a26889f8dc9461aecfa7b7ba13cef0039df3df13: Status 404 returned error can't find the container with id 4440ed44f6d92136c8d6dff7a26889f8dc9461aecfa7b7ba13cef0039df3df13 Dec 02 13:38:45 crc kubenswrapper[4725]: I1202 13:38:45.913836 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx"] Dec 02 13:38:46 crc kubenswrapper[4725]: I1202 13:38:46.623611 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" event={"ID":"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1","Type":"ContainerStarted","Data":"4440ed44f6d92136c8d6dff7a26889f8dc9461aecfa7b7ba13cef0039df3df13"} Dec 02 13:38:47 crc kubenswrapper[4725]: I1202 13:38:47.633402 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" event={"ID":"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1","Type":"ContainerStarted","Data":"59932bfbc86ae9b77d090c810d3a3f4691695644bdc512eacc775807e27ad29a"} Dec 02 13:38:47 crc kubenswrapper[4725]: I1202 13:38:47.658866 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" podStartSLOduration=3.000576772 podStartE2EDuration="3.658838508s" podCreationTimestamp="2025-12-02 13:38:44 +0000 UTC" firstStartedPulling="2025-12-02 13:38:45.909403168 +0000 UTC m=+2056.866044863" lastFinishedPulling="2025-12-02 13:38:46.567664904 +0000 UTC m=+2057.524306599" observedRunningTime="2025-12-02 13:38:47.650733146 +0000 UTC m=+2058.607374851" watchObservedRunningTime="2025-12-02 13:38:47.658838508 +0000 UTC m=+2058.615480203" Dec 02 13:40:24 crc kubenswrapper[4725]: I1202 13:40:24.285077 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:40:24 crc kubenswrapper[4725]: I1202 13:40:24.285608 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:40:54 crc kubenswrapper[4725]: I1202 13:40:54.284250 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:40:54 crc kubenswrapper[4725]: I1202 13:40:54.284901 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.353934 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vkvhz"] Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.359072 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.371859 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkvhz"] Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.534083 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-utilities\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.534178 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-catalog-content\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.534208 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfpj4\" (UniqueName: \"kubernetes.io/projected/7dccb975-c67a-4be1-9ead-558e78dc0db6-kube-api-access-dfpj4\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.636313 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-catalog-content\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.636697 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfpj4\" (UniqueName: \"kubernetes.io/projected/7dccb975-c67a-4be1-9ead-558e78dc0db6-kube-api-access-dfpj4\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.636977 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-utilities\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.637347 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-catalog-content\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.637539 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-utilities\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.663123 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfpj4\" (UniqueName: \"kubernetes.io/projected/7dccb975-c67a-4be1-9ead-558e78dc0db6-kube-api-access-dfpj4\") pod \"certified-operators-vkvhz\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:00 crc kubenswrapper[4725]: I1202 13:41:00.692872 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:01 crc kubenswrapper[4725]: I1202 13:41:01.202656 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkvhz"] Dec 02 13:41:01 crc kubenswrapper[4725]: I1202 13:41:01.866863 4725 generic.go:334] "Generic (PLEG): container finished" podID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerID="fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2" exitCode=0 Dec 02 13:41:01 crc kubenswrapper[4725]: I1202 13:41:01.866914 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerDied","Data":"fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2"} Dec 02 13:41:01 crc kubenswrapper[4725]: I1202 13:41:01.866946 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerStarted","Data":"855b0c036be01187dbd06fe7fa022b5dd9a7b2a092bc59c5f80dc6eca181a5f1"} Dec 02 13:41:01 crc kubenswrapper[4725]: I1202 13:41:01.868573 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:41:02 crc kubenswrapper[4725]: I1202 13:41:02.876766 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerStarted","Data":"34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74"} Dec 02 13:41:03 crc kubenswrapper[4725]: I1202 13:41:03.890582 4725 generic.go:334] "Generic (PLEG): container finished" podID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerID="34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74" exitCode=0 Dec 02 13:41:03 crc kubenswrapper[4725]: I1202 13:41:03.890681 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerDied","Data":"34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74"} Dec 02 13:41:04 crc kubenswrapper[4725]: I1202 13:41:04.904494 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerStarted","Data":"57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a"} Dec 02 13:41:04 crc kubenswrapper[4725]: I1202 13:41:04.926941 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vkvhz" podStartSLOduration=2.229924895 podStartE2EDuration="4.926910542s" podCreationTimestamp="2025-12-02 13:41:00 +0000 UTC" firstStartedPulling="2025-12-02 13:41:01.868348844 +0000 UTC m=+2192.824990539" lastFinishedPulling="2025-12-02 13:41:04.565334491 +0000 UTC m=+2195.521976186" observedRunningTime="2025-12-02 13:41:04.922858021 +0000 UTC m=+2195.879499716" watchObservedRunningTime="2025-12-02 13:41:04.926910542 +0000 UTC m=+2195.883552237" Dec 02 13:41:10 crc kubenswrapper[4725]: I1202 13:41:10.694299 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:10 crc kubenswrapper[4725]: I1202 13:41:10.694898 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:10 crc kubenswrapper[4725]: I1202 13:41:10.741679 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:11 crc kubenswrapper[4725]: I1202 13:41:11.021992 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:11 crc kubenswrapper[4725]: I1202 13:41:11.086585 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkvhz"] Dec 02 13:41:12 crc kubenswrapper[4725]: I1202 13:41:12.979332 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vkvhz" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="registry-server" containerID="cri-o://57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a" gracePeriod=2 Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.550069 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.709840 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfpj4\" (UniqueName: \"kubernetes.io/projected/7dccb975-c67a-4be1-9ead-558e78dc0db6-kube-api-access-dfpj4\") pod \"7dccb975-c67a-4be1-9ead-558e78dc0db6\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.709943 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-catalog-content\") pod \"7dccb975-c67a-4be1-9ead-558e78dc0db6\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.709967 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-utilities\") pod \"7dccb975-c67a-4be1-9ead-558e78dc0db6\" (UID: \"7dccb975-c67a-4be1-9ead-558e78dc0db6\") " Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.710944 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-utilities" (OuterVolumeSpecName: "utilities") pod "7dccb975-c67a-4be1-9ead-558e78dc0db6" (UID: "7dccb975-c67a-4be1-9ead-558e78dc0db6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.716772 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dccb975-c67a-4be1-9ead-558e78dc0db6-kube-api-access-dfpj4" (OuterVolumeSpecName: "kube-api-access-dfpj4") pod "7dccb975-c67a-4be1-9ead-558e78dc0db6" (UID: "7dccb975-c67a-4be1-9ead-558e78dc0db6"). InnerVolumeSpecName "kube-api-access-dfpj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.758063 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7dccb975-c67a-4be1-9ead-558e78dc0db6" (UID: "7dccb975-c67a-4be1-9ead-558e78dc0db6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.811996 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfpj4\" (UniqueName: \"kubernetes.io/projected/7dccb975-c67a-4be1-9ead-558e78dc0db6-kube-api-access-dfpj4\") on node \"crc\" DevicePath \"\"" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.812032 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.812042 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dccb975-c67a-4be1-9ead-558e78dc0db6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.997844 4725 generic.go:334] "Generic (PLEG): container finished" podID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerID="57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a" exitCode=0 Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.997931 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkvhz" Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.997942 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerDied","Data":"57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a"} Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.998013 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkvhz" event={"ID":"7dccb975-c67a-4be1-9ead-558e78dc0db6","Type":"ContainerDied","Data":"855b0c036be01187dbd06fe7fa022b5dd9a7b2a092bc59c5f80dc6eca181a5f1"} Dec 02 13:41:13 crc kubenswrapper[4725]: I1202 13:41:13.998040 4725 scope.go:117] "RemoveContainer" containerID="57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.025332 4725 scope.go:117] "RemoveContainer" containerID="34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.051104 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkvhz"] Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.063026 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vkvhz"] Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.068068 4725 scope.go:117] "RemoveContainer" containerID="fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.120500 4725 scope.go:117] "RemoveContainer" containerID="57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a" Dec 02 13:41:14 crc kubenswrapper[4725]: E1202 13:41:14.121115 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a\": container with ID starting with 57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a not found: ID does not exist" containerID="57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.121171 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a"} err="failed to get container status \"57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a\": rpc error: code = NotFound desc = could not find container \"57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a\": container with ID starting with 57c5ea2646453cc85c2c44afce2c86ae1324f9bec00ad334abac4e24df4fe90a not found: ID does not exist" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.121209 4725 scope.go:117] "RemoveContainer" containerID="34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74" Dec 02 13:41:14 crc kubenswrapper[4725]: E1202 13:41:14.121484 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74\": container with ID starting with 34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74 not found: ID does not exist" containerID="34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.121521 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74"} err="failed to get container status \"34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74\": rpc error: code = NotFound desc = could not find container \"34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74\": container with ID starting with 34ffce4047ee622d7d659b05fd57335a503f7055102647db4ddefe384aabcc74 not found: ID does not exist" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.121538 4725 scope.go:117] "RemoveContainer" containerID="fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2" Dec 02 13:41:14 crc kubenswrapper[4725]: E1202 13:41:14.122353 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2\": container with ID starting with fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2 not found: ID does not exist" containerID="fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2" Dec 02 13:41:14 crc kubenswrapper[4725]: I1202 13:41:14.122518 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2"} err="failed to get container status \"fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2\": rpc error: code = NotFound desc = could not find container \"fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2\": container with ID starting with fa27aeabd5cfc98039fe5b4fb551fe3d83e3bf1392e495786d617815e0857fe2 not found: ID does not exist" Dec 02 13:41:15 crc kubenswrapper[4725]: I1202 13:41:15.282603 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" path="/var/lib/kubelet/pods/7dccb975-c67a-4be1-9ead-558e78dc0db6/volumes" Dec 02 13:41:24 crc kubenswrapper[4725]: I1202 13:41:24.285069 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:41:24 crc kubenswrapper[4725]: I1202 13:41:24.285881 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:41:24 crc kubenswrapper[4725]: I1202 13:41:24.285942 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:41:24 crc kubenswrapper[4725]: I1202 13:41:24.287085 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:41:24 crc kubenswrapper[4725]: I1202 13:41:24.287169 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" gracePeriod=600 Dec 02 13:41:24 crc kubenswrapper[4725]: E1202 13:41:24.448798 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:41:25 crc kubenswrapper[4725]: I1202 13:41:25.110808 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" exitCode=0 Dec 02 13:41:25 crc kubenswrapper[4725]: I1202 13:41:25.110882 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6"} Dec 02 13:41:25 crc kubenswrapper[4725]: I1202 13:41:25.111005 4725 scope.go:117] "RemoveContainer" containerID="dec08dfad0c2e314494f251f9651e69a5944d3ad6d3e3ff5c1d452de251ae268" Dec 02 13:41:25 crc kubenswrapper[4725]: I1202 13:41:25.112434 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:41:25 crc kubenswrapper[4725]: E1202 13:41:25.113356 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:41:36 crc kubenswrapper[4725]: I1202 13:41:36.268497 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:41:36 crc kubenswrapper[4725]: E1202 13:41:36.269180 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:41:50 crc kubenswrapper[4725]: I1202 13:41:50.268965 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:41:50 crc kubenswrapper[4725]: E1202 13:41:50.270476 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:42:02 crc kubenswrapper[4725]: I1202 13:42:02.269035 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:42:02 crc kubenswrapper[4725]: E1202 13:42:02.269882 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:42:16 crc kubenswrapper[4725]: I1202 13:42:16.268318 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:42:16 crc kubenswrapper[4725]: E1202 13:42:16.276762 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:42:29 crc kubenswrapper[4725]: I1202 13:42:29.275910 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:42:29 crc kubenswrapper[4725]: E1202 13:42:29.276943 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:42:41 crc kubenswrapper[4725]: I1202 13:42:41.268147 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:42:41 crc kubenswrapper[4725]: E1202 13:42:41.268944 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:42:41 crc kubenswrapper[4725]: I1202 13:42:41.884700 4725 generic.go:334] "Generic (PLEG): container finished" podID="ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" containerID="59932bfbc86ae9b77d090c810d3a3f4691695644bdc512eacc775807e27ad29a" exitCode=0 Dec 02 13:42:41 crc kubenswrapper[4725]: I1202 13:42:41.884797 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" event={"ID":"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1","Type":"ContainerDied","Data":"59932bfbc86ae9b77d090c810d3a3f4691695644bdc512eacc775807e27ad29a"} Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.350565 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.384804 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qjdv\" (UniqueName: \"kubernetes.io/projected/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-kube-api-access-9qjdv\") pod \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.385128 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-ssh-key\") pod \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.385220 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-inventory\") pod \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.385303 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-secret-0\") pod \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.385347 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-combined-ca-bundle\") pod \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\" (UID: \"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1\") " Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.392003 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" (UID: "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.392732 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-kube-api-access-9qjdv" (OuterVolumeSpecName: "kube-api-access-9qjdv") pod "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" (UID: "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1"). InnerVolumeSpecName "kube-api-access-9qjdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.420238 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" (UID: "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.420761 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-inventory" (OuterVolumeSpecName: "inventory") pod "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" (UID: "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.427726 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" (UID: "ee65f1b9-bbb7-4561-af64-e6f0ab3deec1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.487578 4725 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.488242 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qjdv\" (UniqueName: \"kubernetes.io/projected/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-kube-api-access-9qjdv\") on node \"crc\" DevicePath \"\"" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.488270 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.488280 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.488297 4725 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ee65f1b9-bbb7-4561-af64-e6f0ab3deec1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.901222 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" event={"ID":"ee65f1b9-bbb7-4561-af64-e6f0ab3deec1","Type":"ContainerDied","Data":"4440ed44f6d92136c8d6dff7a26889f8dc9461aecfa7b7ba13cef0039df3df13"} Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.901288 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4440ed44f6d92136c8d6dff7a26889f8dc9461aecfa7b7ba13cef0039df3df13" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.901377 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997051 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc"] Dec 02 13:42:43 crc kubenswrapper[4725]: E1202 13:42:43.997519 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997538 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 13:42:43 crc kubenswrapper[4725]: E1202 13:42:43.997548 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="extract-content" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997556 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="extract-content" Dec 02 13:42:43 crc kubenswrapper[4725]: E1202 13:42:43.997574 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="registry-server" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997581 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="registry-server" Dec 02 13:42:43 crc kubenswrapper[4725]: E1202 13:42:43.997608 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="extract-utilities" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997615 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="extract-utilities" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997857 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dccb975-c67a-4be1-9ead-558e78dc0db6" containerName="registry-server" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.997878 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee65f1b9-bbb7-4561-af64-e6f0ab3deec1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 13:42:43 crc kubenswrapper[4725]: I1202 13:42:43.998662 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.000879 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.000889 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.001012 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.000947 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.001271 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.001352 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.001842 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.007653 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc"] Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.198861 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.198939 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.198983 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.199105 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.199152 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpfq2\" (UniqueName: \"kubernetes.io/projected/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-kube-api-access-cpfq2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.199180 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.199417 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.199546 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.199622 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.300927 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.300982 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301010 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301068 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301097 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpfq2\" (UniqueName: \"kubernetes.io/projected/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-kube-api-access-cpfq2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301116 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301153 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301180 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.301206 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.302684 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.304692 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.305121 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.305333 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.306069 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.307247 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.308245 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.310140 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.323538 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpfq2\" (UniqueName: \"kubernetes.io/projected/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-kube-api-access-cpfq2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-58vqc\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.324187 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.831387 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc"] Dec 02 13:42:44 crc kubenswrapper[4725]: I1202 13:42:44.912321 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" event={"ID":"76bb9f49-7e28-49dc-9946-d3e6de9e6a26","Type":"ContainerStarted","Data":"e299040465429dff7a5f8cf366fae62f2b189cce0600a03f72298b21bd3507cc"} Dec 02 13:42:45 crc kubenswrapper[4725]: I1202 13:42:45.922916 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" event={"ID":"76bb9f49-7e28-49dc-9946-d3e6de9e6a26","Type":"ContainerStarted","Data":"40f2c67efafbe18ade4b1afef8994783256350cb78b945d9d667eb16cf6d3319"} Dec 02 13:42:52 crc kubenswrapper[4725]: I1202 13:42:52.269068 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:42:52 crc kubenswrapper[4725]: E1202 13:42:52.269690 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:43:06 crc kubenswrapper[4725]: I1202 13:43:06.268603 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:43:06 crc kubenswrapper[4725]: E1202 13:43:06.269298 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:43:06 crc kubenswrapper[4725]: I1202 13:43:06.486085 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-6d448f6767-ccmmh" podUID="69be1bb2-b1e2-41fb-991c-c6a12db18fb8" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 02 13:43:17 crc kubenswrapper[4725]: I1202 13:43:17.269573 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:43:17 crc kubenswrapper[4725]: E1202 13:43:17.270405 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:43:31 crc kubenswrapper[4725]: I1202 13:43:31.268189 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:43:31 crc kubenswrapper[4725]: E1202 13:43:31.269348 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:43:42 crc kubenswrapper[4725]: I1202 13:43:42.269264 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:43:42 crc kubenswrapper[4725]: E1202 13:43:42.271607 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:43:56 crc kubenswrapper[4725]: I1202 13:43:56.268579 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:43:56 crc kubenswrapper[4725]: E1202 13:43:56.269335 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.098575 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" podStartSLOduration=86.428933742 podStartE2EDuration="1m27.098555578s" podCreationTimestamp="2025-12-02 13:42:43 +0000 UTC" firstStartedPulling="2025-12-02 13:42:44.832523651 +0000 UTC m=+2295.789165346" lastFinishedPulling="2025-12-02 13:42:45.502145487 +0000 UTC m=+2296.458787182" observedRunningTime="2025-12-02 13:42:45.943957006 +0000 UTC m=+2296.900598721" watchObservedRunningTime="2025-12-02 13:44:10.098555578 +0000 UTC m=+2381.055197273" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.099330 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sxmgc"] Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.101166 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.115074 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sxmgc"] Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.115571 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-catalog-content\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.115686 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrcj\" (UniqueName: \"kubernetes.io/projected/975962cd-d87d-4fa6-b2a8-bb210aabefc5-kube-api-access-slrcj\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.115725 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-utilities\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.217680 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-catalog-content\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.217836 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrcj\" (UniqueName: \"kubernetes.io/projected/975962cd-d87d-4fa6-b2a8-bb210aabefc5-kube-api-access-slrcj\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.217898 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-utilities\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.218415 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-utilities\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.218630 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-catalog-content\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.242721 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrcj\" (UniqueName: \"kubernetes.io/projected/975962cd-d87d-4fa6-b2a8-bb210aabefc5-kube-api-access-slrcj\") pod \"community-operators-sxmgc\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:10 crc kubenswrapper[4725]: I1202 13:44:10.432933 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:11 crc kubenswrapper[4725]: I1202 13:44:11.056408 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sxmgc"] Dec 02 13:44:11 crc kubenswrapper[4725]: I1202 13:44:11.268092 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:44:11 crc kubenswrapper[4725]: E1202 13:44:11.268630 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:44:11 crc kubenswrapper[4725]: I1202 13:44:11.669720 4725 generic.go:334] "Generic (PLEG): container finished" podID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerID="421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416" exitCode=0 Dec 02 13:44:11 crc kubenswrapper[4725]: I1202 13:44:11.669834 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerDied","Data":"421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416"} Dec 02 13:44:11 crc kubenswrapper[4725]: I1202 13:44:11.670083 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerStarted","Data":"40fd838e0bccfdf128d3de403f5c1b701987f25ca13ff44d8bd4233c361d4432"} Dec 02 13:44:14 crc kubenswrapper[4725]: I1202 13:44:14.696870 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerStarted","Data":"632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965"} Dec 02 13:44:15 crc kubenswrapper[4725]: I1202 13:44:15.706960 4725 generic.go:334] "Generic (PLEG): container finished" podID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerID="632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965" exitCode=0 Dec 02 13:44:15 crc kubenswrapper[4725]: I1202 13:44:15.707032 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerDied","Data":"632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965"} Dec 02 13:44:17 crc kubenswrapper[4725]: I1202 13:44:17.739988 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerStarted","Data":"825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834"} Dec 02 13:44:17 crc kubenswrapper[4725]: I1202 13:44:17.765373 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sxmgc" podStartSLOduration=2.589448055 podStartE2EDuration="7.765348863s" podCreationTimestamp="2025-12-02 13:44:10 +0000 UTC" firstStartedPulling="2025-12-02 13:44:11.672457546 +0000 UTC m=+2382.629099241" lastFinishedPulling="2025-12-02 13:44:16.848358354 +0000 UTC m=+2387.805000049" observedRunningTime="2025-12-02 13:44:17.757170019 +0000 UTC m=+2388.713811734" watchObservedRunningTime="2025-12-02 13:44:17.765348863 +0000 UTC m=+2388.721990568" Dec 02 13:44:20 crc kubenswrapper[4725]: I1202 13:44:20.434800 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:20 crc kubenswrapper[4725]: I1202 13:44:20.435357 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:20 crc kubenswrapper[4725]: I1202 13:44:20.479235 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:26 crc kubenswrapper[4725]: I1202 13:44:26.268648 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:44:26 crc kubenswrapper[4725]: E1202 13:44:26.269506 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:44:30 crc kubenswrapper[4725]: I1202 13:44:30.488340 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:30 crc kubenswrapper[4725]: I1202 13:44:30.535095 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sxmgc"] Dec 02 13:44:30 crc kubenswrapper[4725]: I1202 13:44:30.847482 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sxmgc" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="registry-server" containerID="cri-o://825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834" gracePeriod=2 Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.297595 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.434848 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-utilities\") pod \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.434958 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slrcj\" (UniqueName: \"kubernetes.io/projected/975962cd-d87d-4fa6-b2a8-bb210aabefc5-kube-api-access-slrcj\") pod \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.435073 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-catalog-content\") pod \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\" (UID: \"975962cd-d87d-4fa6-b2a8-bb210aabefc5\") " Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.440336 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975962cd-d87d-4fa6-b2a8-bb210aabefc5-kube-api-access-slrcj" (OuterVolumeSpecName: "kube-api-access-slrcj") pod "975962cd-d87d-4fa6-b2a8-bb210aabefc5" (UID: "975962cd-d87d-4fa6-b2a8-bb210aabefc5"). InnerVolumeSpecName "kube-api-access-slrcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.440375 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-utilities" (OuterVolumeSpecName: "utilities") pod "975962cd-d87d-4fa6-b2a8-bb210aabefc5" (UID: "975962cd-d87d-4fa6-b2a8-bb210aabefc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.483780 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "975962cd-d87d-4fa6-b2a8-bb210aabefc5" (UID: "975962cd-d87d-4fa6-b2a8-bb210aabefc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.538065 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.538097 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/975962cd-d87d-4fa6-b2a8-bb210aabefc5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.538111 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slrcj\" (UniqueName: \"kubernetes.io/projected/975962cd-d87d-4fa6-b2a8-bb210aabefc5-kube-api-access-slrcj\") on node \"crc\" DevicePath \"\"" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.862660 4725 generic.go:334] "Generic (PLEG): container finished" podID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerID="825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834" exitCode=0 Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.863006 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerDied","Data":"825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834"} Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.863038 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxmgc" event={"ID":"975962cd-d87d-4fa6-b2a8-bb210aabefc5","Type":"ContainerDied","Data":"40fd838e0bccfdf128d3de403f5c1b701987f25ca13ff44d8bd4233c361d4432"} Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.863057 4725 scope.go:117] "RemoveContainer" containerID="825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.863218 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxmgc" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.924163 4725 scope.go:117] "RemoveContainer" containerID="632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965" Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.924293 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sxmgc"] Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.933295 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sxmgc"] Dec 02 13:44:31 crc kubenswrapper[4725]: I1202 13:44:31.952200 4725 scope.go:117] "RemoveContainer" containerID="421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416" Dec 02 13:44:32 crc kubenswrapper[4725]: I1202 13:44:31.991892 4725 scope.go:117] "RemoveContainer" containerID="825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834" Dec 02 13:44:32 crc kubenswrapper[4725]: E1202 13:44:32.021159 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834\": container with ID starting with 825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834 not found: ID does not exist" containerID="825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834" Dec 02 13:44:32 crc kubenswrapper[4725]: I1202 13:44:32.021232 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834"} err="failed to get container status \"825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834\": rpc error: code = NotFound desc = could not find container \"825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834\": container with ID starting with 825dae4a837569552c2b8d2617e30cc170b228339350f534d47023595cea0834 not found: ID does not exist" Dec 02 13:44:32 crc kubenswrapper[4725]: I1202 13:44:32.021269 4725 scope.go:117] "RemoveContainer" containerID="632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965" Dec 02 13:44:32 crc kubenswrapper[4725]: E1202 13:44:32.022394 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965\": container with ID starting with 632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965 not found: ID does not exist" containerID="632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965" Dec 02 13:44:32 crc kubenswrapper[4725]: I1202 13:44:32.022424 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965"} err="failed to get container status \"632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965\": rpc error: code = NotFound desc = could not find container \"632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965\": container with ID starting with 632f4eb335a2befe20ccf8dc933fa09e8be25234245ca57838af5942c3ad2965 not found: ID does not exist" Dec 02 13:44:32 crc kubenswrapper[4725]: I1202 13:44:32.022440 4725 scope.go:117] "RemoveContainer" containerID="421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416" Dec 02 13:44:32 crc kubenswrapper[4725]: E1202 13:44:32.023010 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416\": container with ID starting with 421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416 not found: ID does not exist" containerID="421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416" Dec 02 13:44:32 crc kubenswrapper[4725]: I1202 13:44:32.023038 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416"} err="failed to get container status \"421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416\": rpc error: code = NotFound desc = could not find container \"421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416\": container with ID starting with 421462de0d4708bf127cbe32096a83876e579cebf308c3f5af0048cd6e5af416 not found: ID does not exist" Dec 02 13:44:33 crc kubenswrapper[4725]: I1202 13:44:33.278315 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" path="/var/lib/kubelet/pods/975962cd-d87d-4fa6-b2a8-bb210aabefc5/volumes" Dec 02 13:44:37 crc kubenswrapper[4725]: I1202 13:44:37.267675 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:44:37 crc kubenswrapper[4725]: E1202 13:44:37.269166 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:44:51 crc kubenswrapper[4725]: I1202 13:44:51.268234 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:44:51 crc kubenswrapper[4725]: E1202 13:44:51.269915 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.147875 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv"] Dec 02 13:45:00 crc kubenswrapper[4725]: E1202 13:45:00.148878 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="extract-utilities" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.148897 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="extract-utilities" Dec 02 13:45:00 crc kubenswrapper[4725]: E1202 13:45:00.148918 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="registry-server" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.148930 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="registry-server" Dec 02 13:45:00 crc kubenswrapper[4725]: E1202 13:45:00.148954 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="extract-content" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.148962 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="extract-content" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.149214 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="975962cd-d87d-4fa6-b2a8-bb210aabefc5" containerName="registry-server" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.150152 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.157855 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.161727 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv"] Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.187545 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.262068 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1af225-5d00-481a-a724-9628a3c309ec-secret-volume\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.262174 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fprgs\" (UniqueName: \"kubernetes.io/projected/6d1af225-5d00-481a-a724-9628a3c309ec-kube-api-access-fprgs\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.262217 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1af225-5d00-481a-a724-9628a3c309ec-config-volume\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.364930 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1af225-5d00-481a-a724-9628a3c309ec-secret-volume\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.364986 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fprgs\" (UniqueName: \"kubernetes.io/projected/6d1af225-5d00-481a-a724-9628a3c309ec-kube-api-access-fprgs\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.365015 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1af225-5d00-481a-a724-9628a3c309ec-config-volume\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.366000 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1af225-5d00-481a-a724-9628a3c309ec-config-volume\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.371759 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1af225-5d00-481a-a724-9628a3c309ec-secret-volume\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.383794 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fprgs\" (UniqueName: \"kubernetes.io/projected/6d1af225-5d00-481a-a724-9628a3c309ec-kube-api-access-fprgs\") pod \"collect-profiles-29411385-ff7tv\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.478169 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:00 crc kubenswrapper[4725]: I1202 13:45:00.949127 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv"] Dec 02 13:45:01 crc kubenswrapper[4725]: I1202 13:45:01.128367 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" event={"ID":"6d1af225-5d00-481a-a724-9628a3c309ec","Type":"ContainerStarted","Data":"a2f24c723f331898e8e2b3176851fee7ad2e11040b22bdd8a42b7f3fb3582dcd"} Dec 02 13:45:02 crc kubenswrapper[4725]: I1202 13:45:02.140804 4725 generic.go:334] "Generic (PLEG): container finished" podID="6d1af225-5d00-481a-a724-9628a3c309ec" containerID="752720e3ec9e32a3dbb7cbd0814f7e183ae49bf522c8513b491d603a98964634" exitCode=0 Dec 02 13:45:02 crc kubenswrapper[4725]: I1202 13:45:02.140876 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" event={"ID":"6d1af225-5d00-481a-a724-9628a3c309ec","Type":"ContainerDied","Data":"752720e3ec9e32a3dbb7cbd0814f7e183ae49bf522c8513b491d603a98964634"} Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.408120 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.431683 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1af225-5d00-481a-a724-9628a3c309ec-config-volume\") pod \"6d1af225-5d00-481a-a724-9628a3c309ec\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.431874 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1af225-5d00-481a-a724-9628a3c309ec-secret-volume\") pod \"6d1af225-5d00-481a-a724-9628a3c309ec\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.432024 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fprgs\" (UniqueName: \"kubernetes.io/projected/6d1af225-5d00-481a-a724-9628a3c309ec-kube-api-access-fprgs\") pod \"6d1af225-5d00-481a-a724-9628a3c309ec\" (UID: \"6d1af225-5d00-481a-a724-9628a3c309ec\") " Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.432482 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1af225-5d00-481a-a724-9628a3c309ec-config-volume" (OuterVolumeSpecName: "config-volume") pod "6d1af225-5d00-481a-a724-9628a3c309ec" (UID: "6d1af225-5d00-481a-a724-9628a3c309ec"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.437531 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1af225-5d00-481a-a724-9628a3c309ec-kube-api-access-fprgs" (OuterVolumeSpecName: "kube-api-access-fprgs") pod "6d1af225-5d00-481a-a724-9628a3c309ec" (UID: "6d1af225-5d00-481a-a724-9628a3c309ec"). InnerVolumeSpecName "kube-api-access-fprgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.437833 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1af225-5d00-481a-a724-9628a3c309ec-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6d1af225-5d00-481a-a724-9628a3c309ec" (UID: "6d1af225-5d00-481a-a724-9628a3c309ec"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.534977 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6d1af225-5d00-481a-a724-9628a3c309ec-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.535007 4725 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6d1af225-5d00-481a-a724-9628a3c309ec-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:03 crc kubenswrapper[4725]: I1202 13:45:03.535017 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fprgs\" (UniqueName: \"kubernetes.io/projected/6d1af225-5d00-481a-a724-9628a3c309ec-kube-api-access-fprgs\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:04 crc kubenswrapper[4725]: I1202 13:45:04.158234 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" event={"ID":"6d1af225-5d00-481a-a724-9628a3c309ec","Type":"ContainerDied","Data":"a2f24c723f331898e8e2b3176851fee7ad2e11040b22bdd8a42b7f3fb3582dcd"} Dec 02 13:45:04 crc kubenswrapper[4725]: I1202 13:45:04.158271 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2f24c723f331898e8e2b3176851fee7ad2e11040b22bdd8a42b7f3fb3582dcd" Dec 02 13:45:04 crc kubenswrapper[4725]: I1202 13:45:04.158295 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411385-ff7tv" Dec 02 13:45:04 crc kubenswrapper[4725]: I1202 13:45:04.501611 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx"] Dec 02 13:45:04 crc kubenswrapper[4725]: I1202 13:45:04.518374 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411340-rnxpx"] Dec 02 13:45:05 crc kubenswrapper[4725]: I1202 13:45:05.281150 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66462d0f-425e-470b-a4e2-94682d534fe9" path="/var/lib/kubelet/pods/66462d0f-425e-470b-a4e2-94682d534fe9/volumes" Dec 02 13:45:06 crc kubenswrapper[4725]: I1202 13:45:06.268587 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:45:06 crc kubenswrapper[4725]: E1202 13:45:06.268955 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:45:17 crc kubenswrapper[4725]: I1202 13:45:17.276631 4725 generic.go:334] "Generic (PLEG): container finished" podID="76bb9f49-7e28-49dc-9946-d3e6de9e6a26" containerID="40f2c67efafbe18ade4b1afef8994783256350cb78b945d9d667eb16cf6d3319" exitCode=0 Dec 02 13:45:17 crc kubenswrapper[4725]: I1202 13:45:17.280924 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" event={"ID":"76bb9f49-7e28-49dc-9946-d3e6de9e6a26","Type":"ContainerDied","Data":"40f2c67efafbe18ade4b1afef8994783256350cb78b945d9d667eb16cf6d3319"} Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.689595 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.834572 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-1\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835054 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-1\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835086 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-ssh-key\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835130 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-0\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835203 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-combined-ca-bundle\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835240 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-inventory\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835260 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-extra-config-0\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835305 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpfq2\" (UniqueName: \"kubernetes.io/projected/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-kube-api-access-cpfq2\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.835327 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-0\") pod \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\" (UID: \"76bb9f49-7e28-49dc-9946-d3e6de9e6a26\") " Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.840627 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-kube-api-access-cpfq2" (OuterVolumeSpecName: "kube-api-access-cpfq2") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "kube-api-access-cpfq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.841957 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.864483 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.864537 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-inventory" (OuterVolumeSpecName: "inventory") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.865688 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.865704 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.866126 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.867710 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.869946 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "76bb9f49-7e28-49dc-9946-d3e6de9e6a26" (UID: "76bb9f49-7e28-49dc-9946-d3e6de9e6a26"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938568 4725 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938618 4725 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938632 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938646 4725 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938655 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpfq2\" (UniqueName: \"kubernetes.io/projected/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-kube-api-access-cpfq2\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938670 4725 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938681 4725 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938695 4725 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:18 crc kubenswrapper[4725]: I1202 13:45:18.938706 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76bb9f49-7e28-49dc-9946-d3e6de9e6a26-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.297483 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" event={"ID":"76bb9f49-7e28-49dc-9946-d3e6de9e6a26","Type":"ContainerDied","Data":"e299040465429dff7a5f8cf366fae62f2b189cce0600a03f72298b21bd3507cc"} Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.297541 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e299040465429dff7a5f8cf366fae62f2b189cce0600a03f72298b21bd3507cc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.297538 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-58vqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.404816 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc"] Dec 02 13:45:19 crc kubenswrapper[4725]: E1202 13:45:19.405300 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76bb9f49-7e28-49dc-9946-d3e6de9e6a26" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.405321 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="76bb9f49-7e28-49dc-9946-d3e6de9e6a26" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 13:45:19 crc kubenswrapper[4725]: E1202 13:45:19.405338 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1af225-5d00-481a-a724-9628a3c309ec" containerName="collect-profiles" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.405347 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1af225-5d00-481a-a724-9628a3c309ec" containerName="collect-profiles" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.405573 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1af225-5d00-481a-a724-9628a3c309ec" containerName="collect-profiles" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.405598 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="76bb9f49-7e28-49dc-9946-d3e6de9e6a26" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.406222 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.411021 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-p7vfs" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.411242 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.411441 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.411749 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.411936 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.423971 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc"] Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550121 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnngz\" (UniqueName: \"kubernetes.io/projected/b45407f7-eb82-40f8-a223-0dfbabc966e5-kube-api-access-qnngz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550200 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550234 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550652 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550715 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550737 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.550839 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.652590 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.652934 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.652965 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.653023 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.653083 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnngz\" (UniqueName: \"kubernetes.io/projected/b45407f7-eb82-40f8-a223-0dfbabc966e5-kube-api-access-qnngz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.653123 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.653145 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.656414 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.657569 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.657650 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.658408 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.659438 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.667675 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.671525 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnngz\" (UniqueName: \"kubernetes.io/projected/b45407f7-eb82-40f8-a223-0dfbabc966e5-kube-api-access-qnngz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:19 crc kubenswrapper[4725]: I1202 13:45:19.731554 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:45:20 crc kubenswrapper[4725]: I1202 13:45:20.242628 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc"] Dec 02 13:45:20 crc kubenswrapper[4725]: I1202 13:45:20.311142 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" event={"ID":"b45407f7-eb82-40f8-a223-0dfbabc966e5","Type":"ContainerStarted","Data":"b52a2e729e8c3c7618b85cdaf4481116bd661e143ca14f5f82414736be396584"} Dec 02 13:45:21 crc kubenswrapper[4725]: I1202 13:45:21.267980 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:45:21 crc kubenswrapper[4725]: E1202 13:45:21.268661 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:45:22 crc kubenswrapper[4725]: I1202 13:45:22.328226 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" event={"ID":"b45407f7-eb82-40f8-a223-0dfbabc966e5","Type":"ContainerStarted","Data":"bf65be4411213397c705da5d760e3218f5758d429545692988e047fd8efc2134"} Dec 02 13:45:22 crc kubenswrapper[4725]: I1202 13:45:22.356155 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" podStartSLOduration=2.497809517 podStartE2EDuration="3.356133456s" podCreationTimestamp="2025-12-02 13:45:19 +0000 UTC" firstStartedPulling="2025-12-02 13:45:20.25200752 +0000 UTC m=+2451.208649215" lastFinishedPulling="2025-12-02 13:45:21.110331459 +0000 UTC m=+2452.066973154" observedRunningTime="2025-12-02 13:45:22.348200338 +0000 UTC m=+2453.304842073" watchObservedRunningTime="2025-12-02 13:45:22.356133456 +0000 UTC m=+2453.312775151" Dec 02 13:45:34 crc kubenswrapper[4725]: I1202 13:45:34.268668 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:45:34 crc kubenswrapper[4725]: E1202 13:45:34.269450 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:45:46 crc kubenswrapper[4725]: I1202 13:45:46.268950 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:45:46 crc kubenswrapper[4725]: E1202 13:45:46.269711 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:45:50 crc kubenswrapper[4725]: I1202 13:45:50.194576 4725 scope.go:117] "RemoveContainer" containerID="3a0e3a76fff0d396e51bf96e1542597b76841b39cadaefda8908c7044594e915" Dec 02 13:45:59 crc kubenswrapper[4725]: I1202 13:45:59.270365 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:45:59 crc kubenswrapper[4725]: E1202 13:45:59.271126 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:46:13 crc kubenswrapper[4725]: I1202 13:46:13.268378 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:46:13 crc kubenswrapper[4725]: E1202 13:46:13.269279 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:46:26 crc kubenswrapper[4725]: I1202 13:46:26.268333 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:46:26 crc kubenswrapper[4725]: I1202 13:46:26.908673 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"6d71ab0b0d5eda10c5632eb9f27e308726113f3a6e14c3263541fab015988e33"} Dec 02 13:47:30 crc kubenswrapper[4725]: I1202 13:47:30.441528 4725 generic.go:334] "Generic (PLEG): container finished" podID="b45407f7-eb82-40f8-a223-0dfbabc966e5" containerID="bf65be4411213397c705da5d760e3218f5758d429545692988e047fd8efc2134" exitCode=0 Dec 02 13:47:30 crc kubenswrapper[4725]: I1202 13:47:30.441603 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" event={"ID":"b45407f7-eb82-40f8-a223-0dfbabc966e5","Type":"ContainerDied","Data":"bf65be4411213397c705da5d760e3218f5758d429545692988e047fd8efc2134"} Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.832913 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.986788 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-telemetry-combined-ca-bundle\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.986871 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ssh-key\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.987062 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-2\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.987091 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-1\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.987197 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-0\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.987269 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnngz\" (UniqueName: \"kubernetes.io/projected/b45407f7-eb82-40f8-a223-0dfbabc966e5-kube-api-access-qnngz\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.987311 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-inventory\") pod \"b45407f7-eb82-40f8-a223-0dfbabc966e5\" (UID: \"b45407f7-eb82-40f8-a223-0dfbabc966e5\") " Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.995238 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:47:31 crc kubenswrapper[4725]: I1202 13:47:31.996738 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b45407f7-eb82-40f8-a223-0dfbabc966e5-kube-api-access-qnngz" (OuterVolumeSpecName: "kube-api-access-qnngz") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "kube-api-access-qnngz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.020390 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.020388 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.022966 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-inventory" (OuterVolumeSpecName: "inventory") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.023081 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.034129 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "b45407f7-eb82-40f8-a223-0dfbabc966e5" (UID: "b45407f7-eb82-40f8-a223-0dfbabc966e5"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089505 4725 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089538 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnngz\" (UniqueName: \"kubernetes.io/projected/b45407f7-eb82-40f8-a223-0dfbabc966e5-kube-api-access-qnngz\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089550 4725 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089561 4725 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089569 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089579 4725 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.089588 4725 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b45407f7-eb82-40f8-a223-0dfbabc966e5-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.463030 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" event={"ID":"b45407f7-eb82-40f8-a223-0dfbabc966e5","Type":"ContainerDied","Data":"b52a2e729e8c3c7618b85cdaf4481116bd661e143ca14f5f82414736be396584"} Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.463083 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b52a2e729e8c3c7618b85cdaf4481116bd661e143ca14f5f82414736be396584" Dec 02 13:47:32 crc kubenswrapper[4725]: I1202 13:47:32.463145 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.025718 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 13:48:33 crc kubenswrapper[4725]: E1202 13:48:33.027088 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b45407f7-eb82-40f8-a223-0dfbabc966e5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.027116 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b45407f7-eb82-40f8-a223-0dfbabc966e5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.027358 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b45407f7-eb82-40f8-a223-0dfbabc966e5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.028012 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.030454 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.030521 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-s6dg7" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.030640 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.030702 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.035579 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048340 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048388 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048431 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048456 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-config-data\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048570 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn5d7\" (UniqueName: \"kubernetes.io/projected/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-kube-api-access-fn5d7\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048593 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048624 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048647 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.048681 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150775 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150829 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150859 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150878 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-config-data\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150955 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn5d7\" (UniqueName: \"kubernetes.io/projected/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-kube-api-access-fn5d7\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150972 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.150993 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.151028 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.151062 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.151185 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.151486 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.152360 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-config-data\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.152655 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.152749 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.157321 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.160199 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.163181 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.171258 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn5d7\" (UniqueName: \"kubernetes.io/projected/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-kube-api-access-fn5d7\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.177342 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.346526 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.782229 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 13:48:33 crc kubenswrapper[4725]: W1202 13:48:33.784348 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod564d9d77_b56c_4e22_a4ae_c6fc6f02e7d8.slice/crio-d95ca1155896a587210cd6e783f0182e0cc449570d6a9f06262b09dd96b7bf1e WatchSource:0}: Error finding container d95ca1155896a587210cd6e783f0182e0cc449570d6a9f06262b09dd96b7bf1e: Status 404 returned error can't find the container with id d95ca1155896a587210cd6e783f0182e0cc449570d6a9f06262b09dd96b7bf1e Dec 02 13:48:33 crc kubenswrapper[4725]: I1202 13:48:33.787084 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:48:34 crc kubenswrapper[4725]: I1202 13:48:34.003132 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8","Type":"ContainerStarted","Data":"d95ca1155896a587210cd6e783f0182e0cc449570d6a9f06262b09dd96b7bf1e"} Dec 02 13:48:41 crc kubenswrapper[4725]: I1202 13:48:41.866509 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lf2fx"] Dec 02 13:48:41 crc kubenswrapper[4725]: I1202 13:48:41.876661 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lf2fx"] Dec 02 13:48:41 crc kubenswrapper[4725]: I1202 13:48:41.876770 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.060294 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-catalog-content\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.060514 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-utilities\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.060587 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpsgt\" (UniqueName: \"kubernetes.io/projected/45ed97fc-71ec-478b-9548-e016bce643eb-kube-api-access-mpsgt\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.163157 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-utilities\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.163241 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpsgt\" (UniqueName: \"kubernetes.io/projected/45ed97fc-71ec-478b-9548-e016bce643eb-kube-api-access-mpsgt\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.163312 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-catalog-content\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.163814 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-utilities\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.163857 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-catalog-content\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.184260 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpsgt\" (UniqueName: \"kubernetes.io/projected/45ed97fc-71ec-478b-9548-e016bce643eb-kube-api-access-mpsgt\") pod \"redhat-operators-lf2fx\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.205557 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:48:42 crc kubenswrapper[4725]: I1202 13:48:42.749544 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lf2fx"] Dec 02 13:48:43 crc kubenswrapper[4725]: I1202 13:48:43.087152 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerStarted","Data":"02e9e057e96942a462bd489651d03fe32aa5d88c46ad83f134ccba3f276f9cdd"} Dec 02 13:48:46 crc kubenswrapper[4725]: I1202 13:48:46.114124 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerStarted","Data":"ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0"} Dec 02 13:48:46 crc kubenswrapper[4725]: E1202 13:48:46.446023 4725 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45ed97fc_71ec_478b_9548_e016bce643eb.slice/crio-conmon-ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0.scope\": RecentStats: unable to find data in memory cache]" Dec 02 13:48:47 crc kubenswrapper[4725]: I1202 13:48:47.130819 4725 generic.go:334] "Generic (PLEG): container finished" podID="45ed97fc-71ec-478b-9548-e016bce643eb" containerID="ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0" exitCode=0 Dec 02 13:48:47 crc kubenswrapper[4725]: I1202 13:48:47.130890 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerDied","Data":"ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0"} Dec 02 13:48:54 crc kubenswrapper[4725]: I1202 13:48:54.284795 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:48:54 crc kubenswrapper[4725]: I1202 13:48:54.285302 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:49:12 crc kubenswrapper[4725]: E1202 13:49:12.015045 4725 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 02 13:49:12 crc kubenswrapper[4725]: E1202 13:49:12.016258 4725 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fn5d7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 13:49:12 crc kubenswrapper[4725]: E1202 13:49:12.017723 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" Dec 02 13:49:12 crc kubenswrapper[4725]: E1202 13:49:12.524064 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" Dec 02 13:49:14 crc kubenswrapper[4725]: I1202 13:49:14.538902 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerStarted","Data":"e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779"} Dec 02 13:49:16 crc kubenswrapper[4725]: I1202 13:49:16.558811 4725 generic.go:334] "Generic (PLEG): container finished" podID="45ed97fc-71ec-478b-9548-e016bce643eb" containerID="e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779" exitCode=0 Dec 02 13:49:16 crc kubenswrapper[4725]: I1202 13:49:16.558896 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerDied","Data":"e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779"} Dec 02 13:49:17 crc kubenswrapper[4725]: I1202 13:49:17.569937 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerStarted","Data":"1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03"} Dec 02 13:49:22 crc kubenswrapper[4725]: I1202 13:49:22.206361 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:49:22 crc kubenswrapper[4725]: I1202 13:49:22.207084 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:49:22 crc kubenswrapper[4725]: I1202 13:49:22.263689 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:49:22 crc kubenswrapper[4725]: I1202 13:49:22.295663 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lf2fx" podStartSLOduration=36.05490596 podStartE2EDuration="41.295637343s" podCreationTimestamp="2025-12-02 13:48:41 +0000 UTC" firstStartedPulling="2025-12-02 13:49:11.892591207 +0000 UTC m=+2682.849232902" lastFinishedPulling="2025-12-02 13:49:17.13332259 +0000 UTC m=+2688.089964285" observedRunningTime="2025-12-02 13:49:17.590529497 +0000 UTC m=+2688.547171212" watchObservedRunningTime="2025-12-02 13:49:22.295637343 +0000 UTC m=+2693.252279038" Dec 02 13:49:22 crc kubenswrapper[4725]: I1202 13:49:22.655196 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:49:22 crc kubenswrapper[4725]: I1202 13:49:22.711216 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lf2fx"] Dec 02 13:49:24 crc kubenswrapper[4725]: I1202 13:49:24.285174 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:49:24 crc kubenswrapper[4725]: I1202 13:49:24.285225 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:49:24 crc kubenswrapper[4725]: I1202 13:49:24.623867 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lf2fx" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="registry-server" containerID="cri-o://1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03" gracePeriod=2 Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.080011 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.268269 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpsgt\" (UniqueName: \"kubernetes.io/projected/45ed97fc-71ec-478b-9548-e016bce643eb-kube-api-access-mpsgt\") pod \"45ed97fc-71ec-478b-9548-e016bce643eb\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.268400 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-catalog-content\") pod \"45ed97fc-71ec-478b-9548-e016bce643eb\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.268523 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-utilities\") pod \"45ed97fc-71ec-478b-9548-e016bce643eb\" (UID: \"45ed97fc-71ec-478b-9548-e016bce643eb\") " Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.269486 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-utilities" (OuterVolumeSpecName: "utilities") pod "45ed97fc-71ec-478b-9548-e016bce643eb" (UID: "45ed97fc-71ec-478b-9548-e016bce643eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.287157 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45ed97fc-71ec-478b-9548-e016bce643eb-kube-api-access-mpsgt" (OuterVolumeSpecName: "kube-api-access-mpsgt") pod "45ed97fc-71ec-478b-9548-e016bce643eb" (UID: "45ed97fc-71ec-478b-9548-e016bce643eb"). InnerVolumeSpecName "kube-api-access-mpsgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.370713 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpsgt\" (UniqueName: \"kubernetes.io/projected/45ed97fc-71ec-478b-9548-e016bce643eb-kube-api-access-mpsgt\") on node \"crc\" DevicePath \"\"" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.370741 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.376525 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45ed97fc-71ec-478b-9548-e016bce643eb" (UID: "45ed97fc-71ec-478b-9548-e016bce643eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.472520 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ed97fc-71ec-478b-9548-e016bce643eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.634621 4725 generic.go:334] "Generic (PLEG): container finished" podID="45ed97fc-71ec-478b-9548-e016bce643eb" containerID="1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03" exitCode=0 Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.634677 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerDied","Data":"1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03"} Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.634683 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf2fx" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.634714 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf2fx" event={"ID":"45ed97fc-71ec-478b-9548-e016bce643eb","Type":"ContainerDied","Data":"02e9e057e96942a462bd489651d03fe32aa5d88c46ad83f134ccba3f276f9cdd"} Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.634736 4725 scope.go:117] "RemoveContainer" containerID="1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.662395 4725 scope.go:117] "RemoveContainer" containerID="e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.680607 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lf2fx"] Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.695446 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lf2fx"] Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.701697 4725 scope.go:117] "RemoveContainer" containerID="ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.767681 4725 scope.go:117] "RemoveContainer" containerID="1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03" Dec 02 13:49:25 crc kubenswrapper[4725]: E1202 13:49:25.768233 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03\": container with ID starting with 1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03 not found: ID does not exist" containerID="1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.768283 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03"} err="failed to get container status \"1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03\": rpc error: code = NotFound desc = could not find container \"1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03\": container with ID starting with 1c86f73e5b5fa437c78bf576858f8e16bdb52e2f952e1fc6799c8826c9881b03 not found: ID does not exist" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.768317 4725 scope.go:117] "RemoveContainer" containerID="e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779" Dec 02 13:49:25 crc kubenswrapper[4725]: E1202 13:49:25.768774 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779\": container with ID starting with e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779 not found: ID does not exist" containerID="e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.768800 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779"} err="failed to get container status \"e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779\": rpc error: code = NotFound desc = could not find container \"e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779\": container with ID starting with e016db454aa7290bb487066bd515e2bbe5683ac76e500d93510301a31146a779 not found: ID does not exist" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.768826 4725 scope.go:117] "RemoveContainer" containerID="ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0" Dec 02 13:49:25 crc kubenswrapper[4725]: E1202 13:49:25.769446 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0\": container with ID starting with ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0 not found: ID does not exist" containerID="ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0" Dec 02 13:49:25 crc kubenswrapper[4725]: I1202 13:49:25.769513 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0"} err="failed to get container status \"ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0\": rpc error: code = NotFound desc = could not find container \"ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0\": container with ID starting with ccdfa924b47ea8e553b65368a06ef141e5e5f24d52f886b7a93efbbbcd8ef5d0 not found: ID does not exist" Dec 02 13:49:27 crc kubenswrapper[4725]: I1202 13:49:27.279003 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" path="/var/lib/kubelet/pods/45ed97fc-71ec-478b-9548-e016bce643eb/volumes" Dec 02 13:49:27 crc kubenswrapper[4725]: I1202 13:49:27.665164 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8","Type":"ContainerStarted","Data":"dc7c71041432622249a52b2983ff9644de358639d489267ccd7465272ae4b34d"} Dec 02 13:49:27 crc kubenswrapper[4725]: I1202 13:49:27.683669 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.701585715 podStartE2EDuration="56.683649916s" podCreationTimestamp="2025-12-02 13:48:31 +0000 UTC" firstStartedPulling="2025-12-02 13:48:33.786862633 +0000 UTC m=+2644.743504328" lastFinishedPulling="2025-12-02 13:49:25.768926834 +0000 UTC m=+2696.725568529" observedRunningTime="2025-12-02 13:49:27.680292744 +0000 UTC m=+2698.636934449" watchObservedRunningTime="2025-12-02 13:49:27.683649916 +0000 UTC m=+2698.640291611" Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.284812 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.285250 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.285297 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.286024 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d71ab0b0d5eda10c5632eb9f27e308726113f3a6e14c3263541fab015988e33"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.286065 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://6d71ab0b0d5eda10c5632eb9f27e308726113f3a6e14c3263541fab015988e33" gracePeriod=600 Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.890302 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="6d71ab0b0d5eda10c5632eb9f27e308726113f3a6e14c3263541fab015988e33" exitCode=0 Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.890373 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"6d71ab0b0d5eda10c5632eb9f27e308726113f3a6e14c3263541fab015988e33"} Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.890814 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1"} Dec 02 13:49:54 crc kubenswrapper[4725]: I1202 13:49:54.890849 4725 scope.go:117] "RemoveContainer" containerID="2a18b94f8df50bb6f67ed7a3fa33a402f1fc7c42e6506bffd7e591daacdf53a6" Dec 02 13:50:51 crc kubenswrapper[4725]: I1202 13:50:51.246832 4725 patch_prober.go:28] interesting pod/controller-manager-7cd8b9fcfc-q5s7l container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 02 13:50:51 crc kubenswrapper[4725]: I1202 13:50:51.247417 4725 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" podUID="befb7e08-5919-4fda-b0fd-e25ff4970f5d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 13:50:51 crc kubenswrapper[4725]: I1202 13:50:51.246962 4725 patch_prober.go:28] interesting pod/controller-manager-7cd8b9fcfc-q5s7l container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 02 13:50:51 crc kubenswrapper[4725]: I1202 13:50:51.247507 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7cd8b9fcfc-q5s7l" podUID="befb7e08-5919-4fda-b0fd-e25ff4970f5d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.099864 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-69jcm"] Dec 02 13:51:18 crc kubenswrapper[4725]: E1202 13:51:18.100848 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="registry-server" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.100862 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="registry-server" Dec 02 13:51:18 crc kubenswrapper[4725]: E1202 13:51:18.100881 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="extract-content" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.100888 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="extract-content" Dec 02 13:51:18 crc kubenswrapper[4725]: E1202 13:51:18.100906 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="extract-utilities" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.100912 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="extract-utilities" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.101187 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="45ed97fc-71ec-478b-9548-e016bce643eb" containerName="registry-server" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.102664 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.110457 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-69jcm"] Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.218965 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsb7r\" (UniqueName: \"kubernetes.io/projected/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-kube-api-access-gsb7r\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.219090 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-utilities\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.219139 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-catalog-content\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.320614 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsb7r\" (UniqueName: \"kubernetes.io/projected/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-kube-api-access-gsb7r\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.320717 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-utilities\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.320753 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-catalog-content\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.321337 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-utilities\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.321423 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-catalog-content\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.342688 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsb7r\" (UniqueName: \"kubernetes.io/projected/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-kube-api-access-gsb7r\") pod \"certified-operators-69jcm\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:18 crc kubenswrapper[4725]: I1202 13:51:18.446636 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:19 crc kubenswrapper[4725]: I1202 13:51:19.014275 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-69jcm"] Dec 02 13:51:19 crc kubenswrapper[4725]: I1202 13:51:19.662523 4725 generic.go:334] "Generic (PLEG): container finished" podID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerID="e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5" exitCode=0 Dec 02 13:51:19 crc kubenswrapper[4725]: I1202 13:51:19.662590 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69jcm" event={"ID":"25b7dde7-a903-45fa-9449-9a0b0a83b4b5","Type":"ContainerDied","Data":"e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5"} Dec 02 13:51:19 crc kubenswrapper[4725]: I1202 13:51:19.662853 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69jcm" event={"ID":"25b7dde7-a903-45fa-9449-9a0b0a83b4b5","Type":"ContainerStarted","Data":"55da59339e3082d678398f987cbb4b320672e62726ec4bfa73dfc0116e4c52e1"} Dec 02 13:51:21 crc kubenswrapper[4725]: I1202 13:51:21.681683 4725 generic.go:334] "Generic (PLEG): container finished" podID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerID="8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f" exitCode=0 Dec 02 13:51:21 crc kubenswrapper[4725]: I1202 13:51:21.681782 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69jcm" event={"ID":"25b7dde7-a903-45fa-9449-9a0b0a83b4b5","Type":"ContainerDied","Data":"8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f"} Dec 02 13:51:22 crc kubenswrapper[4725]: I1202 13:51:22.692050 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69jcm" event={"ID":"25b7dde7-a903-45fa-9449-9a0b0a83b4b5","Type":"ContainerStarted","Data":"e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784"} Dec 02 13:51:22 crc kubenswrapper[4725]: I1202 13:51:22.710792 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-69jcm" podStartSLOduration=2.288951677 podStartE2EDuration="4.710769927s" podCreationTimestamp="2025-12-02 13:51:18 +0000 UTC" firstStartedPulling="2025-12-02 13:51:19.664584005 +0000 UTC m=+2810.621225700" lastFinishedPulling="2025-12-02 13:51:22.086402255 +0000 UTC m=+2813.043043950" observedRunningTime="2025-12-02 13:51:22.708189073 +0000 UTC m=+2813.664830768" watchObservedRunningTime="2025-12-02 13:51:22.710769927 +0000 UTC m=+2813.667411622" Dec 02 13:51:28 crc kubenswrapper[4725]: I1202 13:51:28.446943 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:28 crc kubenswrapper[4725]: I1202 13:51:28.447652 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:28 crc kubenswrapper[4725]: I1202 13:51:28.489918 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:28 crc kubenswrapper[4725]: I1202 13:51:28.804614 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:28 crc kubenswrapper[4725]: I1202 13:51:28.851445 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-69jcm"] Dec 02 13:51:30 crc kubenswrapper[4725]: I1202 13:51:30.770063 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-69jcm" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="registry-server" containerID="cri-o://e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784" gracePeriod=2 Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.271189 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.381352 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsb7r\" (UniqueName: \"kubernetes.io/projected/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-kube-api-access-gsb7r\") pod \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.381753 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-utilities\") pod \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.381799 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-catalog-content\") pod \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\" (UID: \"25b7dde7-a903-45fa-9449-9a0b0a83b4b5\") " Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.383536 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-utilities" (OuterVolumeSpecName: "utilities") pod "25b7dde7-a903-45fa-9449-9a0b0a83b4b5" (UID: "25b7dde7-a903-45fa-9449-9a0b0a83b4b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.391636 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-kube-api-access-gsb7r" (OuterVolumeSpecName: "kube-api-access-gsb7r") pod "25b7dde7-a903-45fa-9449-9a0b0a83b4b5" (UID: "25b7dde7-a903-45fa-9449-9a0b0a83b4b5"). InnerVolumeSpecName "kube-api-access-gsb7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.433247 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25b7dde7-a903-45fa-9449-9a0b0a83b4b5" (UID: "25b7dde7-a903-45fa-9449-9a0b0a83b4b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.484025 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.484063 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.484081 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsb7r\" (UniqueName: \"kubernetes.io/projected/25b7dde7-a903-45fa-9449-9a0b0a83b4b5-kube-api-access-gsb7r\") on node \"crc\" DevicePath \"\"" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.780211 4725 generic.go:334] "Generic (PLEG): container finished" podID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerID="e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784" exitCode=0 Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.780278 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69jcm" event={"ID":"25b7dde7-a903-45fa-9449-9a0b0a83b4b5","Type":"ContainerDied","Data":"e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784"} Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.780300 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-69jcm" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.780326 4725 scope.go:117] "RemoveContainer" containerID="e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.780313 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-69jcm" event={"ID":"25b7dde7-a903-45fa-9449-9a0b0a83b4b5","Type":"ContainerDied","Data":"55da59339e3082d678398f987cbb4b320672e62726ec4bfa73dfc0116e4c52e1"} Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.803816 4725 scope.go:117] "RemoveContainer" containerID="8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.812567 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-69jcm"] Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.822919 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-69jcm"] Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.840530 4725 scope.go:117] "RemoveContainer" containerID="e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.867269 4725 scope.go:117] "RemoveContainer" containerID="e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784" Dec 02 13:51:31 crc kubenswrapper[4725]: E1202 13:51:31.867894 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784\": container with ID starting with e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784 not found: ID does not exist" containerID="e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.867928 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784"} err="failed to get container status \"e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784\": rpc error: code = NotFound desc = could not find container \"e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784\": container with ID starting with e6136dadc824315aa70f3bfe99d68df90a8f3d5bb6a78b0d09b071437bfe3784 not found: ID does not exist" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.867951 4725 scope.go:117] "RemoveContainer" containerID="8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f" Dec 02 13:51:31 crc kubenswrapper[4725]: E1202 13:51:31.868405 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f\": container with ID starting with 8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f not found: ID does not exist" containerID="8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.868428 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f"} err="failed to get container status \"8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f\": rpc error: code = NotFound desc = could not find container \"8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f\": container with ID starting with 8cb00fb92dd9c9f652ea4705f4afabd597e11ccf3e047e84ff2588563be9c67f not found: ID does not exist" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.868441 4725 scope.go:117] "RemoveContainer" containerID="e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5" Dec 02 13:51:31 crc kubenswrapper[4725]: E1202 13:51:31.868795 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5\": container with ID starting with e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5 not found: ID does not exist" containerID="e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5" Dec 02 13:51:31 crc kubenswrapper[4725]: I1202 13:51:31.868815 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5"} err="failed to get container status \"e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5\": rpc error: code = NotFound desc = could not find container \"e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5\": container with ID starting with e15e018a5cb7dc03e77a2518bc4f89546ee1f5bd5e641ef1b8f9e79e75463fe5 not found: ID does not exist" Dec 02 13:51:33 crc kubenswrapper[4725]: I1202 13:51:33.281987 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" path="/var/lib/kubelet/pods/25b7dde7-a903-45fa-9449-9a0b0a83b4b5/volumes" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.172490 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-krt5m"] Dec 02 13:51:39 crc kubenswrapper[4725]: E1202 13:51:39.173676 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="extract-content" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.173692 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="extract-content" Dec 02 13:51:39 crc kubenswrapper[4725]: E1202 13:51:39.173708 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="extract-utilities" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.173716 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="extract-utilities" Dec 02 13:51:39 crc kubenswrapper[4725]: E1202 13:51:39.173728 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="registry-server" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.173736 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="registry-server" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.173987 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b7dde7-a903-45fa-9449-9a0b0a83b4b5" containerName="registry-server" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.176665 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.190436 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krt5m"] Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.338616 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-catalog-content\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.338699 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4tz4\" (UniqueName: \"kubernetes.io/projected/b0d66770-00ad-4194-8927-6b709151af87-kube-api-access-g4tz4\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.338913 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-utilities\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.440358 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-utilities\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.440522 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-catalog-content\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.440565 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4tz4\" (UniqueName: \"kubernetes.io/projected/b0d66770-00ad-4194-8927-6b709151af87-kube-api-access-g4tz4\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.440935 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-utilities\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.441060 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-catalog-content\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.460608 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4tz4\" (UniqueName: \"kubernetes.io/projected/b0d66770-00ad-4194-8927-6b709151af87-kube-api-access-g4tz4\") pod \"redhat-marketplace-krt5m\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.498301 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:39 crc kubenswrapper[4725]: I1202 13:51:39.988871 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krt5m"] Dec 02 13:51:40 crc kubenswrapper[4725]: I1202 13:51:40.860516 4725 generic.go:334] "Generic (PLEG): container finished" podID="b0d66770-00ad-4194-8927-6b709151af87" containerID="21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b" exitCode=0 Dec 02 13:51:40 crc kubenswrapper[4725]: I1202 13:51:40.860668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krt5m" event={"ID":"b0d66770-00ad-4194-8927-6b709151af87","Type":"ContainerDied","Data":"21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b"} Dec 02 13:51:40 crc kubenswrapper[4725]: I1202 13:51:40.861570 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krt5m" event={"ID":"b0d66770-00ad-4194-8927-6b709151af87","Type":"ContainerStarted","Data":"4df6ae8d73e9231f1af56b4f06db2292e96df0f4ed2f8b5173df58d81aa4eb43"} Dec 02 13:51:42 crc kubenswrapper[4725]: I1202 13:51:42.887961 4725 generic.go:334] "Generic (PLEG): container finished" podID="b0d66770-00ad-4194-8927-6b709151af87" containerID="6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9" exitCode=0 Dec 02 13:51:42 crc kubenswrapper[4725]: I1202 13:51:42.888092 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krt5m" event={"ID":"b0d66770-00ad-4194-8927-6b709151af87","Type":"ContainerDied","Data":"6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9"} Dec 02 13:51:43 crc kubenswrapper[4725]: I1202 13:51:43.900409 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krt5m" event={"ID":"b0d66770-00ad-4194-8927-6b709151af87","Type":"ContainerStarted","Data":"a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0"} Dec 02 13:51:43 crc kubenswrapper[4725]: I1202 13:51:43.931000 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-krt5m" podStartSLOduration=2.4842581519999998 podStartE2EDuration="4.930978702s" podCreationTimestamp="2025-12-02 13:51:39 +0000 UTC" firstStartedPulling="2025-12-02 13:51:40.865809658 +0000 UTC m=+2831.822451353" lastFinishedPulling="2025-12-02 13:51:43.312530208 +0000 UTC m=+2834.269171903" observedRunningTime="2025-12-02 13:51:43.928732037 +0000 UTC m=+2834.885373732" watchObservedRunningTime="2025-12-02 13:51:43.930978702 +0000 UTC m=+2834.887620397" Dec 02 13:51:49 crc kubenswrapper[4725]: I1202 13:51:49.498628 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:49 crc kubenswrapper[4725]: I1202 13:51:49.499195 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:49 crc kubenswrapper[4725]: I1202 13:51:49.555493 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:49 crc kubenswrapper[4725]: I1202 13:51:49.995643 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:50 crc kubenswrapper[4725]: I1202 13:51:50.056310 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-krt5m"] Dec 02 13:51:51 crc kubenswrapper[4725]: I1202 13:51:51.964803 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-krt5m" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="registry-server" containerID="cri-o://a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0" gracePeriod=2 Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.498983 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.614166 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-catalog-content\") pod \"b0d66770-00ad-4194-8927-6b709151af87\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.614293 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-utilities\") pod \"b0d66770-00ad-4194-8927-6b709151af87\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.614622 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4tz4\" (UniqueName: \"kubernetes.io/projected/b0d66770-00ad-4194-8927-6b709151af87-kube-api-access-g4tz4\") pod \"b0d66770-00ad-4194-8927-6b709151af87\" (UID: \"b0d66770-00ad-4194-8927-6b709151af87\") " Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.616738 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-utilities" (OuterVolumeSpecName: "utilities") pod "b0d66770-00ad-4194-8927-6b709151af87" (UID: "b0d66770-00ad-4194-8927-6b709151af87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.622119 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0d66770-00ad-4194-8927-6b709151af87-kube-api-access-g4tz4" (OuterVolumeSpecName: "kube-api-access-g4tz4") pod "b0d66770-00ad-4194-8927-6b709151af87" (UID: "b0d66770-00ad-4194-8927-6b709151af87"). InnerVolumeSpecName "kube-api-access-g4tz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.634716 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0d66770-00ad-4194-8927-6b709151af87" (UID: "b0d66770-00ad-4194-8927-6b709151af87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.716901 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.717415 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4tz4\" (UniqueName: \"kubernetes.io/projected/b0d66770-00ad-4194-8927-6b709151af87-kube-api-access-g4tz4\") on node \"crc\" DevicePath \"\"" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.717492 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0d66770-00ad-4194-8927-6b709151af87-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.975837 4725 generic.go:334] "Generic (PLEG): container finished" podID="b0d66770-00ad-4194-8927-6b709151af87" containerID="a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0" exitCode=0 Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.975903 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krt5m" event={"ID":"b0d66770-00ad-4194-8927-6b709151af87","Type":"ContainerDied","Data":"a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0"} Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.975939 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krt5m" event={"ID":"b0d66770-00ad-4194-8927-6b709151af87","Type":"ContainerDied","Data":"4df6ae8d73e9231f1af56b4f06db2292e96df0f4ed2f8b5173df58d81aa4eb43"} Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.975967 4725 scope.go:117] "RemoveContainer" containerID="a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.975989 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krt5m" Dec 02 13:51:52 crc kubenswrapper[4725]: I1202 13:51:52.998847 4725 scope.go:117] "RemoveContainer" containerID="6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.020182 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-krt5m"] Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.024582 4725 scope.go:117] "RemoveContainer" containerID="21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.027989 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-krt5m"] Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.080281 4725 scope.go:117] "RemoveContainer" containerID="a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0" Dec 02 13:51:53 crc kubenswrapper[4725]: E1202 13:51:53.084276 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0\": container with ID starting with a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0 not found: ID does not exist" containerID="a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.084877 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0"} err="failed to get container status \"a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0\": rpc error: code = NotFound desc = could not find container \"a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0\": container with ID starting with a3290a41658b3fab437ed74d24df5da985a001b36199b52621d766d452f70be0 not found: ID does not exist" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.085069 4725 scope.go:117] "RemoveContainer" containerID="6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9" Dec 02 13:51:53 crc kubenswrapper[4725]: E1202 13:51:53.085610 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9\": container with ID starting with 6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9 not found: ID does not exist" containerID="6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.085752 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9"} err="failed to get container status \"6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9\": rpc error: code = NotFound desc = could not find container \"6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9\": container with ID starting with 6d73fc5b19ab4d1d064eb663d95f1cfc2f1bb20b48170eb3bfcd89542fb4b8b9 not found: ID does not exist" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.085864 4725 scope.go:117] "RemoveContainer" containerID="21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b" Dec 02 13:51:53 crc kubenswrapper[4725]: E1202 13:51:53.086203 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b\": container with ID starting with 21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b not found: ID does not exist" containerID="21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.086315 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b"} err="failed to get container status \"21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b\": rpc error: code = NotFound desc = could not find container \"21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b\": container with ID starting with 21ffd675e527c08c37b1a56777855c54427e69f7dbcf60eed3b1a107af341d7b not found: ID does not exist" Dec 02 13:51:53 crc kubenswrapper[4725]: I1202 13:51:53.280843 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0d66770-00ad-4194-8927-6b709151af87" path="/var/lib/kubelet/pods/b0d66770-00ad-4194-8927-6b709151af87/volumes" Dec 02 13:51:54 crc kubenswrapper[4725]: I1202 13:51:54.284900 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:51:54 crc kubenswrapper[4725]: I1202 13:51:54.285598 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:52:24 crc kubenswrapper[4725]: I1202 13:52:24.284509 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:52:24 crc kubenswrapper[4725]: I1202 13:52:24.285050 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.284902 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.285571 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.285622 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.286360 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.286415 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" gracePeriod=600 Dec 02 13:52:54 crc kubenswrapper[4725]: E1202 13:52:54.411829 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.526264 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" exitCode=0 Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.526311 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1"} Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.526357 4725 scope.go:117] "RemoveContainer" containerID="6d71ab0b0d5eda10c5632eb9f27e308726113f3a6e14c3263541fab015988e33" Dec 02 13:52:54 crc kubenswrapper[4725]: I1202 13:52:54.527757 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:52:54 crc kubenswrapper[4725]: E1202 13:52:54.528011 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:53:09 crc kubenswrapper[4725]: I1202 13:53:09.273985 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:53:09 crc kubenswrapper[4725]: E1202 13:53:09.274739 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:53:24 crc kubenswrapper[4725]: I1202 13:53:24.269194 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:53:24 crc kubenswrapper[4725]: E1202 13:53:24.270149 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:53:35 crc kubenswrapper[4725]: I1202 13:53:35.268273 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:53:35 crc kubenswrapper[4725]: E1202 13:53:35.269004 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:53:48 crc kubenswrapper[4725]: I1202 13:53:48.268237 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:53:48 crc kubenswrapper[4725]: E1202 13:53:48.269062 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:54:00 crc kubenswrapper[4725]: I1202 13:54:00.268449 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:54:00 crc kubenswrapper[4725]: E1202 13:54:00.269192 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:54:12 crc kubenswrapper[4725]: I1202 13:54:12.268954 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:54:12 crc kubenswrapper[4725]: E1202 13:54:12.269615 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:54:23 crc kubenswrapper[4725]: I1202 13:54:23.269064 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:54:23 crc kubenswrapper[4725]: E1202 13:54:23.269892 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:54:34 crc kubenswrapper[4725]: I1202 13:54:34.269233 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:54:34 crc kubenswrapper[4725]: E1202 13:54:34.270736 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:54:45 crc kubenswrapper[4725]: I1202 13:54:45.268724 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:54:45 crc kubenswrapper[4725]: E1202 13:54:45.270507 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:54:56 crc kubenswrapper[4725]: I1202 13:54:56.268507 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:54:56 crc kubenswrapper[4725]: E1202 13:54:56.269292 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:55:09 crc kubenswrapper[4725]: I1202 13:55:09.274892 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:55:09 crc kubenswrapper[4725]: E1202 13:55:09.275811 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.288115 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4bbzs"] Dec 02 13:55:22 crc kubenswrapper[4725]: E1202 13:55:22.289200 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="registry-server" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.289216 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="registry-server" Dec 02 13:55:22 crc kubenswrapper[4725]: E1202 13:55:22.289253 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="extract-utilities" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.289260 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="extract-utilities" Dec 02 13:55:22 crc kubenswrapper[4725]: E1202 13:55:22.289277 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="extract-content" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.289283 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="extract-content" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.289475 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0d66770-00ad-4194-8927-6b709151af87" containerName="registry-server" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.291212 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.303612 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4bbzs"] Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.343563 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx5jj\" (UniqueName: \"kubernetes.io/projected/7e90cda4-2698-4e36-b9fc-83bab6844463-kube-api-access-kx5jj\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.343967 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-catalog-content\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.344139 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-utilities\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.445664 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-utilities\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.445813 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx5jj\" (UniqueName: \"kubernetes.io/projected/7e90cda4-2698-4e36-b9fc-83bab6844463-kube-api-access-kx5jj\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.445903 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-catalog-content\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.446282 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-utilities\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.446337 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-catalog-content\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.470603 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx5jj\" (UniqueName: \"kubernetes.io/projected/7e90cda4-2698-4e36-b9fc-83bab6844463-kube-api-access-kx5jj\") pod \"community-operators-4bbzs\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:22 crc kubenswrapper[4725]: I1202 13:55:22.617976 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:23 crc kubenswrapper[4725]: I1202 13:55:23.168498 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4bbzs"] Dec 02 13:55:23 crc kubenswrapper[4725]: I1202 13:55:23.269700 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:55:23 crc kubenswrapper[4725]: E1202 13:55:23.270031 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:55:23 crc kubenswrapper[4725]: I1202 13:55:23.864251 4725 generic.go:334] "Generic (PLEG): container finished" podID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerID="9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652" exitCode=0 Dec 02 13:55:23 crc kubenswrapper[4725]: I1202 13:55:23.864305 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerDied","Data":"9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652"} Dec 02 13:55:23 crc kubenswrapper[4725]: I1202 13:55:23.864398 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerStarted","Data":"7c4f0bce82613ad0651f551b608d2b441aeb7a8032483f6148dbd12f35e02416"} Dec 02 13:55:23 crc kubenswrapper[4725]: I1202 13:55:23.868059 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 13:55:24 crc kubenswrapper[4725]: I1202 13:55:24.893967 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerStarted","Data":"25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb"} Dec 02 13:55:26 crc kubenswrapper[4725]: I1202 13:55:26.913970 4725 generic.go:334] "Generic (PLEG): container finished" podID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerID="25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb" exitCode=0 Dec 02 13:55:26 crc kubenswrapper[4725]: I1202 13:55:26.914062 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerDied","Data":"25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb"} Dec 02 13:55:28 crc kubenswrapper[4725]: I1202 13:55:28.931081 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerStarted","Data":"77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9"} Dec 02 13:55:28 crc kubenswrapper[4725]: I1202 13:55:28.958414 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4bbzs" podStartSLOduration=2.967422373 podStartE2EDuration="6.958390398s" podCreationTimestamp="2025-12-02 13:55:22 +0000 UTC" firstStartedPulling="2025-12-02 13:55:23.867830061 +0000 UTC m=+3054.824471756" lastFinishedPulling="2025-12-02 13:55:27.858798086 +0000 UTC m=+3058.815439781" observedRunningTime="2025-12-02 13:55:28.949101597 +0000 UTC m=+3059.905743292" watchObservedRunningTime="2025-12-02 13:55:28.958390398 +0000 UTC m=+3059.915032093" Dec 02 13:55:32 crc kubenswrapper[4725]: I1202 13:55:32.619053 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:32 crc kubenswrapper[4725]: I1202 13:55:32.621741 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:32 crc kubenswrapper[4725]: I1202 13:55:32.663801 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:34 crc kubenswrapper[4725]: I1202 13:55:34.022323 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:34 crc kubenswrapper[4725]: I1202 13:55:34.070098 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4bbzs"] Dec 02 13:55:35 crc kubenswrapper[4725]: I1202 13:55:35.993815 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4bbzs" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="registry-server" containerID="cri-o://77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9" gracePeriod=2 Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.494293 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.532637 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-utilities\") pod \"7e90cda4-2698-4e36-b9fc-83bab6844463\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.532707 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx5jj\" (UniqueName: \"kubernetes.io/projected/7e90cda4-2698-4e36-b9fc-83bab6844463-kube-api-access-kx5jj\") pod \"7e90cda4-2698-4e36-b9fc-83bab6844463\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.533258 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-utilities" (OuterVolumeSpecName: "utilities") pod "7e90cda4-2698-4e36-b9fc-83bab6844463" (UID: "7e90cda4-2698-4e36-b9fc-83bab6844463"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.533288 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-catalog-content\") pod \"7e90cda4-2698-4e36-b9fc-83bab6844463\" (UID: \"7e90cda4-2698-4e36-b9fc-83bab6844463\") " Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.534121 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.542208 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e90cda4-2698-4e36-b9fc-83bab6844463-kube-api-access-kx5jj" (OuterVolumeSpecName: "kube-api-access-kx5jj") pod "7e90cda4-2698-4e36-b9fc-83bab6844463" (UID: "7e90cda4-2698-4e36-b9fc-83bab6844463"). InnerVolumeSpecName "kube-api-access-kx5jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.599775 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e90cda4-2698-4e36-b9fc-83bab6844463" (UID: "7e90cda4-2698-4e36-b9fc-83bab6844463"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.636163 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx5jj\" (UniqueName: \"kubernetes.io/projected/7e90cda4-2698-4e36-b9fc-83bab6844463-kube-api-access-kx5jj\") on node \"crc\" DevicePath \"\"" Dec 02 13:55:36 crc kubenswrapper[4725]: I1202 13:55:36.636214 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90cda4-2698-4e36-b9fc-83bab6844463-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.006306 4725 generic.go:334] "Generic (PLEG): container finished" podID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerID="77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9" exitCode=0 Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.006441 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerDied","Data":"77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9"} Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.006615 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbzs" event={"ID":"7e90cda4-2698-4e36-b9fc-83bab6844463","Type":"ContainerDied","Data":"7c4f0bce82613ad0651f551b608d2b441aeb7a8032483f6148dbd12f35e02416"} Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.006640 4725 scope.go:117] "RemoveContainer" containerID="77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.006512 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbzs" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.033988 4725 scope.go:117] "RemoveContainer" containerID="25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.046166 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4bbzs"] Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.069754 4725 scope.go:117] "RemoveContainer" containerID="9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.084803 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4bbzs"] Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.116114 4725 scope.go:117] "RemoveContainer" containerID="77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9" Dec 02 13:55:37 crc kubenswrapper[4725]: E1202 13:55:37.116951 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9\": container with ID starting with 77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9 not found: ID does not exist" containerID="77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.117118 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9"} err="failed to get container status \"77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9\": rpc error: code = NotFound desc = could not find container \"77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9\": container with ID starting with 77f178c8be0fae31e4d1fcaf22796059e4736230ddd463c33446715b53f9d1e9 not found: ID does not exist" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.117157 4725 scope.go:117] "RemoveContainer" containerID="25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb" Dec 02 13:55:37 crc kubenswrapper[4725]: E1202 13:55:37.117633 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb\": container with ID starting with 25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb not found: ID does not exist" containerID="25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.117759 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb"} err="failed to get container status \"25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb\": rpc error: code = NotFound desc = could not find container \"25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb\": container with ID starting with 25cd4544adab86b8f09144f5b153c86457670e3689f114c9546d8fda90e445fb not found: ID does not exist" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.117844 4725 scope.go:117] "RemoveContainer" containerID="9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652" Dec 02 13:55:37 crc kubenswrapper[4725]: E1202 13:55:37.118639 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652\": container with ID starting with 9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652 not found: ID does not exist" containerID="9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.118672 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652"} err="failed to get container status \"9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652\": rpc error: code = NotFound desc = could not find container \"9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652\": container with ID starting with 9dde3b2769a72545e72ea7c4a8a656ffa5c8a540dd269f3e283d60be644f7652 not found: ID does not exist" Dec 02 13:55:37 crc kubenswrapper[4725]: I1202 13:55:37.279381 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" path="/var/lib/kubelet/pods/7e90cda4-2698-4e36-b9fc-83bab6844463/volumes" Dec 02 13:55:38 crc kubenswrapper[4725]: I1202 13:55:38.269190 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:55:38 crc kubenswrapper[4725]: E1202 13:55:38.269971 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:55:49 crc kubenswrapper[4725]: I1202 13:55:49.276072 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:55:49 crc kubenswrapper[4725]: E1202 13:55:49.277016 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:56:03 crc kubenswrapper[4725]: I1202 13:56:03.268872 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:56:03 crc kubenswrapper[4725]: E1202 13:56:03.269715 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:56:14 crc kubenswrapper[4725]: I1202 13:56:14.268941 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:56:14 crc kubenswrapper[4725]: E1202 13:56:14.269881 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:56:28 crc kubenswrapper[4725]: I1202 13:56:28.268214 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:56:28 crc kubenswrapper[4725]: E1202 13:56:28.268955 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:56:41 crc kubenswrapper[4725]: I1202 13:56:41.267827 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:56:41 crc kubenswrapper[4725]: E1202 13:56:41.268421 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:56:53 crc kubenswrapper[4725]: I1202 13:56:53.268258 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:56:53 crc kubenswrapper[4725]: E1202 13:56:53.269116 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:57:04 crc kubenswrapper[4725]: I1202 13:57:04.268672 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:57:04 crc kubenswrapper[4725]: E1202 13:57:04.269480 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:57:18 crc kubenswrapper[4725]: I1202 13:57:18.267790 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:57:18 crc kubenswrapper[4725]: E1202 13:57:18.269875 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:57:31 crc kubenswrapper[4725]: I1202 13:57:31.268748 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:57:31 crc kubenswrapper[4725]: E1202 13:57:31.269727 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:57:42 crc kubenswrapper[4725]: I1202 13:57:42.268408 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:57:42 crc kubenswrapper[4725]: E1202 13:57:42.269199 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 13:57:57 crc kubenswrapper[4725]: I1202 13:57:57.268352 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 13:57:58 crc kubenswrapper[4725]: I1202 13:57:58.167988 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"cadf99b0b053173fd0df960b6f5ab3a95c2f07c85dde7cc800424b5829a75d8b"} Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.169641 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-92bkd"] Dec 02 13:59:01 crc kubenswrapper[4725]: E1202 13:59:01.170726 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="extract-utilities" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.170742 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="extract-utilities" Dec 02 13:59:01 crc kubenswrapper[4725]: E1202 13:59:01.170771 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="registry-server" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.170779 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="registry-server" Dec 02 13:59:01 crc kubenswrapper[4725]: E1202 13:59:01.170810 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="extract-content" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.170817 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="extract-content" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.171133 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e90cda4-2698-4e36-b9fc-83bab6844463" containerName="registry-server" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.172901 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.191427 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-92bkd"] Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.261544 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-catalog-content\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.261603 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-utilities\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.261650 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp795\" (UniqueName: \"kubernetes.io/projected/a63727c6-916e-4844-bb2d-d9d5d267d663-kube-api-access-lp795\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.364738 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-catalog-content\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.364856 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-utilities\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.364915 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp795\" (UniqueName: \"kubernetes.io/projected/a63727c6-916e-4844-bb2d-d9d5d267d663-kube-api-access-lp795\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.367059 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-catalog-content\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.367070 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-utilities\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.405284 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp795\" (UniqueName: \"kubernetes.io/projected/a63727c6-916e-4844-bb2d-d9d5d267d663-kube-api-access-lp795\") pod \"redhat-operators-92bkd\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:01 crc kubenswrapper[4725]: I1202 13:59:01.498728 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:02 crc kubenswrapper[4725]: I1202 13:59:02.062110 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-92bkd"] Dec 02 13:59:02 crc kubenswrapper[4725]: I1202 13:59:02.747509 4725 generic.go:334] "Generic (PLEG): container finished" podID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerID="37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4" exitCode=0 Dec 02 13:59:02 crc kubenswrapper[4725]: I1202 13:59:02.747602 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerDied","Data":"37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4"} Dec 02 13:59:02 crc kubenswrapper[4725]: I1202 13:59:02.747896 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerStarted","Data":"c53b7dffd2ff55256426db63d21509e13329e1c561ed8698c3ef1f26961d6520"} Dec 02 13:59:04 crc kubenswrapper[4725]: I1202 13:59:04.766970 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerStarted","Data":"88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018"} Dec 02 13:59:06 crc kubenswrapper[4725]: I1202 13:59:06.788838 4725 generic.go:334] "Generic (PLEG): container finished" podID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerID="88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018" exitCode=0 Dec 02 13:59:06 crc kubenswrapper[4725]: I1202 13:59:06.788938 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerDied","Data":"88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018"} Dec 02 13:59:08 crc kubenswrapper[4725]: I1202 13:59:08.810842 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerStarted","Data":"28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3"} Dec 02 13:59:08 crc kubenswrapper[4725]: I1202 13:59:08.840151 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-92bkd" podStartSLOduration=2.825312902 podStartE2EDuration="7.840131923s" podCreationTimestamp="2025-12-02 13:59:01 +0000 UTC" firstStartedPulling="2025-12-02 13:59:02.749732948 +0000 UTC m=+3273.706374643" lastFinishedPulling="2025-12-02 13:59:07.764551959 +0000 UTC m=+3278.721193664" observedRunningTime="2025-12-02 13:59:08.83199115 +0000 UTC m=+3279.788632845" watchObservedRunningTime="2025-12-02 13:59:08.840131923 +0000 UTC m=+3279.796773618" Dec 02 13:59:11 crc kubenswrapper[4725]: I1202 13:59:11.499837 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:11 crc kubenswrapper[4725]: I1202 13:59:11.500208 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:12 crc kubenswrapper[4725]: I1202 13:59:12.544343 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-92bkd" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="registry-server" probeResult="failure" output=< Dec 02 13:59:12 crc kubenswrapper[4725]: timeout: failed to connect service ":50051" within 1s Dec 02 13:59:12 crc kubenswrapper[4725]: > Dec 02 13:59:21 crc kubenswrapper[4725]: I1202 13:59:21.550600 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:21 crc kubenswrapper[4725]: I1202 13:59:21.603755 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:21 crc kubenswrapper[4725]: I1202 13:59:21.791400 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-92bkd"] Dec 02 13:59:22 crc kubenswrapper[4725]: I1202 13:59:22.950354 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-92bkd" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="registry-server" containerID="cri-o://28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3" gracePeriod=2 Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.476737 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.530627 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-catalog-content\") pod \"a63727c6-916e-4844-bb2d-d9d5d267d663\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.530698 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp795\" (UniqueName: \"kubernetes.io/projected/a63727c6-916e-4844-bb2d-d9d5d267d663-kube-api-access-lp795\") pod \"a63727c6-916e-4844-bb2d-d9d5d267d663\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.530869 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-utilities\") pod \"a63727c6-916e-4844-bb2d-d9d5d267d663\" (UID: \"a63727c6-916e-4844-bb2d-d9d5d267d663\") " Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.531695 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-utilities" (OuterVolumeSpecName: "utilities") pod "a63727c6-916e-4844-bb2d-d9d5d267d663" (UID: "a63727c6-916e-4844-bb2d-d9d5d267d663"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.536912 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63727c6-916e-4844-bb2d-d9d5d267d663-kube-api-access-lp795" (OuterVolumeSpecName: "kube-api-access-lp795") pod "a63727c6-916e-4844-bb2d-d9d5d267d663" (UID: "a63727c6-916e-4844-bb2d-d9d5d267d663"). InnerVolumeSpecName "kube-api-access-lp795". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.632355 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp795\" (UniqueName: \"kubernetes.io/projected/a63727c6-916e-4844-bb2d-d9d5d267d663-kube-api-access-lp795\") on node \"crc\" DevicePath \"\"" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.632407 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.659734 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a63727c6-916e-4844-bb2d-d9d5d267d663" (UID: "a63727c6-916e-4844-bb2d-d9d5d267d663"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.734635 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a63727c6-916e-4844-bb2d-d9d5d267d663-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.961502 4725 generic.go:334] "Generic (PLEG): container finished" podID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerID="28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3" exitCode=0 Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.961550 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92bkd" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.961547 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerDied","Data":"28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3"} Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.961668 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92bkd" event={"ID":"a63727c6-916e-4844-bb2d-d9d5d267d663","Type":"ContainerDied","Data":"c53b7dffd2ff55256426db63d21509e13329e1c561ed8698c3ef1f26961d6520"} Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.961699 4725 scope.go:117] "RemoveContainer" containerID="28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3" Dec 02 13:59:23 crc kubenswrapper[4725]: I1202 13:59:23.987755 4725 scope.go:117] "RemoveContainer" containerID="88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.015336 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-92bkd"] Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.029019 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-92bkd"] Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.033703 4725 scope.go:117] "RemoveContainer" containerID="37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.060341 4725 scope.go:117] "RemoveContainer" containerID="28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3" Dec 02 13:59:24 crc kubenswrapper[4725]: E1202 13:59:24.060861 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3\": container with ID starting with 28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3 not found: ID does not exist" containerID="28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.060903 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3"} err="failed to get container status \"28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3\": rpc error: code = NotFound desc = could not find container \"28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3\": container with ID starting with 28bd580bec878c078c94a6e559f011e437291ef222dbdd1db974d78107f665e3 not found: ID does not exist" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.060927 4725 scope.go:117] "RemoveContainer" containerID="88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018" Dec 02 13:59:24 crc kubenswrapper[4725]: E1202 13:59:24.061257 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018\": container with ID starting with 88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018 not found: ID does not exist" containerID="88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.061304 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018"} err="failed to get container status \"88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018\": rpc error: code = NotFound desc = could not find container \"88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018\": container with ID starting with 88049bc9981c71fcc3371a8965b72eb986cc22695702c90a8908f71ef472b018 not found: ID does not exist" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.061338 4725 scope.go:117] "RemoveContainer" containerID="37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4" Dec 02 13:59:24 crc kubenswrapper[4725]: E1202 13:59:24.061776 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4\": container with ID starting with 37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4 not found: ID does not exist" containerID="37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4" Dec 02 13:59:24 crc kubenswrapper[4725]: I1202 13:59:24.061815 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4"} err="failed to get container status \"37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4\": rpc error: code = NotFound desc = could not find container \"37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4\": container with ID starting with 37ae9666b940c3464ebc469f47ecd1933730f822177175479136994f6f9a10f4 not found: ID does not exist" Dec 02 13:59:25 crc kubenswrapper[4725]: I1202 13:59:25.281015 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" path="/var/lib/kubelet/pods/a63727c6-916e-4844-bb2d-d9d5d267d663/volumes" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.144303 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp"] Dec 02 14:00:00 crc kubenswrapper[4725]: E1202 14:00:00.145282 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="extract-content" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.145298 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="extract-content" Dec 02 14:00:00 crc kubenswrapper[4725]: E1202 14:00:00.145326 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="extract-utilities" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.145333 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="extract-utilities" Dec 02 14:00:00 crc kubenswrapper[4725]: E1202 14:00:00.145349 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="registry-server" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.145357 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="registry-server" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.145605 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63727c6-916e-4844-bb2d-d9d5d267d663" containerName="registry-server" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.146293 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.148196 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.148405 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.172683 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp"] Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.241338 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/286719af-be17-4913-83cf-4f551c5a0b7c-secret-volume\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.241499 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/286719af-be17-4913-83cf-4f551c5a0b7c-config-volume\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.241563 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh4wx\" (UniqueName: \"kubernetes.io/projected/286719af-be17-4913-83cf-4f551c5a0b7c-kube-api-access-vh4wx\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.343330 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/286719af-be17-4913-83cf-4f551c5a0b7c-config-volume\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.343489 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh4wx\" (UniqueName: \"kubernetes.io/projected/286719af-be17-4913-83cf-4f551c5a0b7c-kube-api-access-vh4wx\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.343711 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/286719af-be17-4913-83cf-4f551c5a0b7c-secret-volume\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.344373 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/286719af-be17-4913-83cf-4f551c5a0b7c-config-volume\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.350622 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/286719af-be17-4913-83cf-4f551c5a0b7c-secret-volume\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.361826 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh4wx\" (UniqueName: \"kubernetes.io/projected/286719af-be17-4913-83cf-4f551c5a0b7c-kube-api-access-vh4wx\") pod \"collect-profiles-29411400-xttvp\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.474058 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:00 crc kubenswrapper[4725]: I1202 14:00:00.912730 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp"] Dec 02 14:00:01 crc kubenswrapper[4725]: I1202 14:00:01.407642 4725 generic.go:334] "Generic (PLEG): container finished" podID="286719af-be17-4913-83cf-4f551c5a0b7c" containerID="bfff221fe12d6a2ead087d45519e21c110c2240dc8b43ef22689770ee8c87ec9" exitCode=0 Dec 02 14:00:01 crc kubenswrapper[4725]: I1202 14:00:01.407779 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" event={"ID":"286719af-be17-4913-83cf-4f551c5a0b7c","Type":"ContainerDied","Data":"bfff221fe12d6a2ead087d45519e21c110c2240dc8b43ef22689770ee8c87ec9"} Dec 02 14:00:01 crc kubenswrapper[4725]: I1202 14:00:01.407990 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" event={"ID":"286719af-be17-4913-83cf-4f551c5a0b7c","Type":"ContainerStarted","Data":"599339ee97ce1479e30bb1000865550fe9cc8531d1588420c007b4de84755e1d"} Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.834945 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.893090 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh4wx\" (UniqueName: \"kubernetes.io/projected/286719af-be17-4913-83cf-4f551c5a0b7c-kube-api-access-vh4wx\") pod \"286719af-be17-4913-83cf-4f551c5a0b7c\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.893178 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/286719af-be17-4913-83cf-4f551c5a0b7c-config-volume\") pod \"286719af-be17-4913-83cf-4f551c5a0b7c\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.893430 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/286719af-be17-4913-83cf-4f551c5a0b7c-secret-volume\") pod \"286719af-be17-4913-83cf-4f551c5a0b7c\" (UID: \"286719af-be17-4913-83cf-4f551c5a0b7c\") " Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.894113 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286719af-be17-4913-83cf-4f551c5a0b7c-config-volume" (OuterVolumeSpecName: "config-volume") pod "286719af-be17-4913-83cf-4f551c5a0b7c" (UID: "286719af-be17-4913-83cf-4f551c5a0b7c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.899055 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286719af-be17-4913-83cf-4f551c5a0b7c-kube-api-access-vh4wx" (OuterVolumeSpecName: "kube-api-access-vh4wx") pod "286719af-be17-4913-83cf-4f551c5a0b7c" (UID: "286719af-be17-4913-83cf-4f551c5a0b7c"). InnerVolumeSpecName "kube-api-access-vh4wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.899500 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286719af-be17-4913-83cf-4f551c5a0b7c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "286719af-be17-4913-83cf-4f551c5a0b7c" (UID: "286719af-be17-4913-83cf-4f551c5a0b7c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.995528 4725 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/286719af-be17-4913-83cf-4f551c5a0b7c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.995885 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh4wx\" (UniqueName: \"kubernetes.io/projected/286719af-be17-4913-83cf-4f551c5a0b7c-kube-api-access-vh4wx\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:02 crc kubenswrapper[4725]: I1202 14:00:02.995897 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/286719af-be17-4913-83cf-4f551c5a0b7c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:03 crc kubenswrapper[4725]: E1202 14:00:03.341075 4725 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod286719af_be17_4913_83cf_4f551c5a0b7c.slice\": RecentStats: unable to find data in memory cache]" Dec 02 14:00:03 crc kubenswrapper[4725]: I1202 14:00:03.425152 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" event={"ID":"286719af-be17-4913-83cf-4f551c5a0b7c","Type":"ContainerDied","Data":"599339ee97ce1479e30bb1000865550fe9cc8531d1588420c007b4de84755e1d"} Dec 02 14:00:03 crc kubenswrapper[4725]: I1202 14:00:03.425202 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="599339ee97ce1479e30bb1000865550fe9cc8531d1588420c007b4de84755e1d" Dec 02 14:00:03 crc kubenswrapper[4725]: I1202 14:00:03.425225 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411400-xttvp" Dec 02 14:00:03 crc kubenswrapper[4725]: I1202 14:00:03.916345 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl"] Dec 02 14:00:03 crc kubenswrapper[4725]: I1202 14:00:03.928555 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411355-vhtcl"] Dec 02 14:00:05 crc kubenswrapper[4725]: I1202 14:00:05.281320 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80abbd6-c57b-484e-9d79-c0809affe488" path="/var/lib/kubelet/pods/b80abbd6-c57b-484e-9d79-c0809affe488/volumes" Dec 02 14:00:20 crc kubenswrapper[4725]: I1202 14:00:20.637840 4725 generic.go:334] "Generic (PLEG): container finished" podID="564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" containerID="dc7c71041432622249a52b2983ff9644de358639d489267ccd7465272ae4b34d" exitCode=0 Dec 02 14:00:20 crc kubenswrapper[4725]: I1202 14:00:20.637936 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8","Type":"ContainerDied","Data":"dc7c71041432622249a52b2983ff9644de358639d489267ccd7465272ae4b34d"} Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.031302 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.145949 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ssh-key\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146029 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146068 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146132 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-workdir\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146222 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-temporary\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146258 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn5d7\" (UniqueName: \"kubernetes.io/projected/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-kube-api-access-fn5d7\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146284 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ca-certs\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146424 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config-secret\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.146452 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-config-data\") pod \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\" (UID: \"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8\") " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.149640 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-config-data" (OuterVolumeSpecName: "config-data") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.154801 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.155743 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.156265 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.161709 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-kube-api-access-fn5d7" (OuterVolumeSpecName: "kube-api-access-fn5d7") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "kube-api-access-fn5d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.182496 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.185679 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.189342 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.209943 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" (UID: "564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250538 4725 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250587 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn5d7\" (UniqueName: \"kubernetes.io/projected/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-kube-api-access-fn5d7\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250600 4725 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250609 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250618 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250628 4725 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250662 4725 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250672 4725 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.250683 4725 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.271290 4725 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.352419 4725 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.658979 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8","Type":"ContainerDied","Data":"d95ca1155896a587210cd6e783f0182e0cc449570d6a9f06262b09dd96b7bf1e"} Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.659034 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d95ca1155896a587210cd6e783f0182e0cc449570d6a9f06262b09dd96b7bf1e" Dec 02 14:00:22 crc kubenswrapper[4725]: I1202 14:00:22.659088 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 14:00:24 crc kubenswrapper[4725]: I1202 14:00:24.284552 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:00:24 crc kubenswrapper[4725]: I1202 14:00:24.285107 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.050790 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 14:00:27 crc kubenswrapper[4725]: E1202 14:00:27.051695 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286719af-be17-4913-83cf-4f551c5a0b7c" containerName="collect-profiles" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.051717 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="286719af-be17-4913-83cf-4f551c5a0b7c" containerName="collect-profiles" Dec 02 14:00:27 crc kubenswrapper[4725]: E1202 14:00:27.051740 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" containerName="tempest-tests-tempest-tests-runner" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.051749 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" containerName="tempest-tests-tempest-tests-runner" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.051993 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="286719af-be17-4913-83cf-4f551c5a0b7c" containerName="collect-profiles" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.052009 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8" containerName="tempest-tests-tempest-tests-runner" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.052836 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.063092 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.125002 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-s6dg7" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.253241 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.253503 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7rp9\" (UniqueName: \"kubernetes.io/projected/32b05d70-2f30-48dd-8fba-8187969352da-kube-api-access-p7rp9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.354829 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7rp9\" (UniqueName: \"kubernetes.io/projected/32b05d70-2f30-48dd-8fba-8187969352da-kube-api-access-p7rp9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.355287 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.355688 4725 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.378301 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7rp9\" (UniqueName: \"kubernetes.io/projected/32b05d70-2f30-48dd-8fba-8187969352da-kube-api-access-p7rp9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.387800 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"32b05d70-2f30-48dd-8fba-8187969352da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.439861 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.876290 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 14:00:27 crc kubenswrapper[4725]: I1202 14:00:27.882233 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 14:00:28 crc kubenswrapper[4725]: I1202 14:00:28.719253 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"32b05d70-2f30-48dd-8fba-8187969352da","Type":"ContainerStarted","Data":"202150b73efcfb39af16f7bb72dc1e79765abd06a9e4151c7bd2e82f70c40213"} Dec 02 14:00:29 crc kubenswrapper[4725]: I1202 14:00:29.733634 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"32b05d70-2f30-48dd-8fba-8187969352da","Type":"ContainerStarted","Data":"20dde07361149dcea6670b69874272de40740e9cf94c57318d78ffe511ddf9e5"} Dec 02 14:00:50 crc kubenswrapper[4725]: I1202 14:00:50.739092 4725 scope.go:117] "RemoveContainer" containerID="33e6c40f5fd039ed85bfa4f7d396640baa09eae2c4aed6594a7d1d1eb957b7a8" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.465852 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=25.422318874 podStartE2EDuration="26.46582345s" podCreationTimestamp="2025-12-02 14:00:27 +0000 UTC" firstStartedPulling="2025-12-02 14:00:27.882004327 +0000 UTC m=+3358.838646022" lastFinishedPulling="2025-12-02 14:00:28.925508903 +0000 UTC m=+3359.882150598" observedRunningTime="2025-12-02 14:00:29.747427393 +0000 UTC m=+3360.704069088" watchObservedRunningTime="2025-12-02 14:00:53.46582345 +0000 UTC m=+3384.422465145" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.468125 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jn99z/must-gather-dpsd4"] Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.470225 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.474206 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jn99z"/"default-dockercfg-x8mj7" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.474912 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jn99z"/"kube-root-ca.crt" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.475930 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jn99z"/"openshift-service-ca.crt" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.486492 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jn99z/must-gather-dpsd4"] Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.587220 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/54433b1a-0157-4690-bffb-01bb356b64fc-must-gather-output\") pod \"must-gather-dpsd4\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.587308 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mq22\" (UniqueName: \"kubernetes.io/projected/54433b1a-0157-4690-bffb-01bb356b64fc-kube-api-access-5mq22\") pod \"must-gather-dpsd4\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.689743 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/54433b1a-0157-4690-bffb-01bb356b64fc-must-gather-output\") pod \"must-gather-dpsd4\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.689782 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mq22\" (UniqueName: \"kubernetes.io/projected/54433b1a-0157-4690-bffb-01bb356b64fc-kube-api-access-5mq22\") pod \"must-gather-dpsd4\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.690255 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/54433b1a-0157-4690-bffb-01bb356b64fc-must-gather-output\") pod \"must-gather-dpsd4\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.708961 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mq22\" (UniqueName: \"kubernetes.io/projected/54433b1a-0157-4690-bffb-01bb356b64fc-kube-api-access-5mq22\") pod \"must-gather-dpsd4\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:53 crc kubenswrapper[4725]: I1202 14:00:53.789172 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:00:54 crc kubenswrapper[4725]: I1202 14:00:54.247030 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jn99z/must-gather-dpsd4"] Dec 02 14:00:54 crc kubenswrapper[4725]: I1202 14:00:54.284241 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:00:54 crc kubenswrapper[4725]: I1202 14:00:54.284304 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:00:54 crc kubenswrapper[4725]: I1202 14:00:54.955828 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/must-gather-dpsd4" event={"ID":"54433b1a-0157-4690-bffb-01bb356b64fc","Type":"ContainerStarted","Data":"78403c94d126371565b5d6c920c8f17a19c59334e7eb5cba7e4b0b264265f967"} Dec 02 14:00:58 crc kubenswrapper[4725]: I1202 14:00:58.989980 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/must-gather-dpsd4" event={"ID":"54433b1a-0157-4690-bffb-01bb356b64fc","Type":"ContainerStarted","Data":"49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a"} Dec 02 14:00:58 crc kubenswrapper[4725]: I1202 14:00:58.992801 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/must-gather-dpsd4" event={"ID":"54433b1a-0157-4690-bffb-01bb356b64fc","Type":"ContainerStarted","Data":"90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7"} Dec 02 14:00:59 crc kubenswrapper[4725]: I1202 14:00:59.027575 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jn99z/must-gather-dpsd4" podStartSLOduration=2.196364337 podStartE2EDuration="6.027539155s" podCreationTimestamp="2025-12-02 14:00:53 +0000 UTC" firstStartedPulling="2025-12-02 14:00:54.254291737 +0000 UTC m=+3385.210933432" lastFinishedPulling="2025-12-02 14:00:58.085466555 +0000 UTC m=+3389.042108250" observedRunningTime="2025-12-02 14:00:59.017582457 +0000 UTC m=+3389.974224152" watchObservedRunningTime="2025-12-02 14:00:59.027539155 +0000 UTC m=+3389.984180870" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.157288 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29411401-76w49"] Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.159235 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.176048 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29411401-76w49"] Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.332319 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-combined-ca-bundle\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.332374 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-config-data\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.332447 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-fernet-keys\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.332587 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld8hb\" (UniqueName: \"kubernetes.io/projected/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-kube-api-access-ld8hb\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.435088 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-combined-ca-bundle\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.435528 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-config-data\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.435658 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-fernet-keys\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.436265 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld8hb\" (UniqueName: \"kubernetes.io/projected/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-kube-api-access-ld8hb\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.442384 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-combined-ca-bundle\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.443974 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-config-data\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.451256 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-fernet-keys\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.454643 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld8hb\" (UniqueName: \"kubernetes.io/projected/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-kube-api-access-ld8hb\") pod \"keystone-cron-29411401-76w49\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:00 crc kubenswrapper[4725]: I1202 14:01:00.481374 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:01 crc kubenswrapper[4725]: I1202 14:01:00.999507 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29411401-76w49"] Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.018430 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411401-76w49" event={"ID":"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70","Type":"ContainerStarted","Data":"dd951f9ef9aafcd6650cb61f416d68e882aa3565ff1cd96fa6f7f84b365c8207"} Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.019049 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411401-76w49" event={"ID":"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70","Type":"ContainerStarted","Data":"41a48f41bff64b053ab88dda98afb162813d4b6388c26b562fd73569280df292"} Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.039369 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29411401-76w49" podStartSLOduration=2.039348777 podStartE2EDuration="2.039348777s" podCreationTimestamp="2025-12-02 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 14:01:02.031917012 +0000 UTC m=+3392.988558707" watchObservedRunningTime="2025-12-02 14:01:02.039348777 +0000 UTC m=+3392.995990472" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.366413 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jn99z/crc-debug-v7gqc"] Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.367882 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.518507 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grxnl\" (UniqueName: \"kubernetes.io/projected/ed2871c7-0ac2-42bf-9e37-3960588f4448-kube-api-access-grxnl\") pod \"crc-debug-v7gqc\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.519220 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed2871c7-0ac2-42bf-9e37-3960588f4448-host\") pod \"crc-debug-v7gqc\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.621479 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grxnl\" (UniqueName: \"kubernetes.io/projected/ed2871c7-0ac2-42bf-9e37-3960588f4448-kube-api-access-grxnl\") pod \"crc-debug-v7gqc\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.621897 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed2871c7-0ac2-42bf-9e37-3960588f4448-host\") pod \"crc-debug-v7gqc\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.622018 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed2871c7-0ac2-42bf-9e37-3960588f4448-host\") pod \"crc-debug-v7gqc\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.656341 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grxnl\" (UniqueName: \"kubernetes.io/projected/ed2871c7-0ac2-42bf-9e37-3960588f4448-kube-api-access-grxnl\") pod \"crc-debug-v7gqc\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:02 crc kubenswrapper[4725]: I1202 14:01:02.689530 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:01:03 crc kubenswrapper[4725]: I1202 14:01:03.030545 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" event={"ID":"ed2871c7-0ac2-42bf-9e37-3960588f4448","Type":"ContainerStarted","Data":"cd0e2b613f975dd96be2a1bb752ff7ed82308ecca4ea149ec9196d1ecc787697"} Dec 02 14:01:04 crc kubenswrapper[4725]: I1202 14:01:04.039899 4725 generic.go:334] "Generic (PLEG): container finished" podID="7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" containerID="dd951f9ef9aafcd6650cb61f416d68e882aa3565ff1cd96fa6f7f84b365c8207" exitCode=0 Dec 02 14:01:04 crc kubenswrapper[4725]: I1202 14:01:04.039943 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411401-76w49" event={"ID":"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70","Type":"ContainerDied","Data":"dd951f9ef9aafcd6650cb61f416d68e882aa3565ff1cd96fa6f7f84b365c8207"} Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.441085 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.575645 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-combined-ca-bundle\") pod \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.575903 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld8hb\" (UniqueName: \"kubernetes.io/projected/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-kube-api-access-ld8hb\") pod \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.575958 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-config-data\") pod \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.577921 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-fernet-keys\") pod \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\" (UID: \"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70\") " Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.586804 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-kube-api-access-ld8hb" (OuterVolumeSpecName: "kube-api-access-ld8hb") pod "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" (UID: "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70"). InnerVolumeSpecName "kube-api-access-ld8hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.598973 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" (UID: "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.616443 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" (UID: "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.657661 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-config-data" (OuterVolumeSpecName: "config-data") pod "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" (UID: "7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.682188 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld8hb\" (UniqueName: \"kubernetes.io/projected/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-kube-api-access-ld8hb\") on node \"crc\" DevicePath \"\"" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.682239 4725 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.682256 4725 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 14:01:05 crc kubenswrapper[4725]: I1202 14:01:05.682267 4725 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 14:01:06 crc kubenswrapper[4725]: I1202 14:01:06.057674 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29411401-76w49" event={"ID":"7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70","Type":"ContainerDied","Data":"41a48f41bff64b053ab88dda98afb162813d4b6388c26b562fd73569280df292"} Dec 02 14:01:06 crc kubenswrapper[4725]: I1202 14:01:06.057998 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41a48f41bff64b053ab88dda98afb162813d4b6388c26b562fd73569280df292" Dec 02 14:01:06 crc kubenswrapper[4725]: I1202 14:01:06.057846 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29411401-76w49" Dec 02 14:01:17 crc kubenswrapper[4725]: I1202 14:01:17.166042 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" event={"ID":"ed2871c7-0ac2-42bf-9e37-3960588f4448","Type":"ContainerStarted","Data":"e652ccab5f70d4e41f7ce29e2c2f39d201dfcee921585830836390b17b685db4"} Dec 02 14:01:17 crc kubenswrapper[4725]: I1202 14:01:17.192286 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" podStartSLOduration=1.271600919 podStartE2EDuration="15.192267889s" podCreationTimestamp="2025-12-02 14:01:02 +0000 UTC" firstStartedPulling="2025-12-02 14:01:02.738804467 +0000 UTC m=+3393.695446152" lastFinishedPulling="2025-12-02 14:01:16.659471427 +0000 UTC m=+3407.616113122" observedRunningTime="2025-12-02 14:01:17.184859945 +0000 UTC m=+3408.141501660" watchObservedRunningTime="2025-12-02 14:01:17.192267889 +0000 UTC m=+3408.148909574" Dec 02 14:01:24 crc kubenswrapper[4725]: I1202 14:01:24.284711 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:01:24 crc kubenswrapper[4725]: I1202 14:01:24.285300 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:01:24 crc kubenswrapper[4725]: I1202 14:01:24.285358 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 14:01:24 crc kubenswrapper[4725]: I1202 14:01:24.286334 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cadf99b0b053173fd0df960b6f5ab3a95c2f07c85dde7cc800424b5829a75d8b"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 14:01:24 crc kubenswrapper[4725]: I1202 14:01:24.286392 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://cadf99b0b053173fd0df960b6f5ab3a95c2f07c85dde7cc800424b5829a75d8b" gracePeriod=600 Dec 02 14:01:25 crc kubenswrapper[4725]: I1202 14:01:25.239027 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="cadf99b0b053173fd0df960b6f5ab3a95c2f07c85dde7cc800424b5829a75d8b" exitCode=0 Dec 02 14:01:25 crc kubenswrapper[4725]: I1202 14:01:25.239093 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"cadf99b0b053173fd0df960b6f5ab3a95c2f07c85dde7cc800424b5829a75d8b"} Dec 02 14:01:25 crc kubenswrapper[4725]: I1202 14:01:25.239417 4725 scope.go:117] "RemoveContainer" containerID="5c50280aa0a26cedc75795337c3c604067baa68570f9686e0b46ac419b0af6c1" Dec 02 14:01:26 crc kubenswrapper[4725]: I1202 14:01:26.251560 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376"} Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.472963 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wnm28"] Dec 02 14:01:44 crc kubenswrapper[4725]: E1202 14:01:44.474090 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" containerName="keystone-cron" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.474109 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" containerName="keystone-cron" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.474370 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70" containerName="keystone-cron" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.481139 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.496293 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wnm28"] Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.638499 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqln6\" (UniqueName: \"kubernetes.io/projected/d8fd3136-ffdb-41b9-b49c-fce1777f4374-kube-api-access-sqln6\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.638582 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-utilities\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.638653 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-catalog-content\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.740866 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqln6\" (UniqueName: \"kubernetes.io/projected/d8fd3136-ffdb-41b9-b49c-fce1777f4374-kube-api-access-sqln6\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.740958 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-utilities\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.741019 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-catalog-content\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.741749 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-utilities\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.741874 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-catalog-content\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.765698 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqln6\" (UniqueName: \"kubernetes.io/projected/d8fd3136-ffdb-41b9-b49c-fce1777f4374-kube-api-access-sqln6\") pod \"certified-operators-wnm28\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:44 crc kubenswrapper[4725]: I1202 14:01:44.804253 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:01:45 crc kubenswrapper[4725]: I1202 14:01:45.406682 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wnm28"] Dec 02 14:01:45 crc kubenswrapper[4725]: I1202 14:01:45.426066 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnm28" event={"ID":"d8fd3136-ffdb-41b9-b49c-fce1777f4374","Type":"ContainerStarted","Data":"495bd1d671bb796f172e64bc5b9fb62af6d02acec0eb1f25a8801fc5524062e6"} Dec 02 14:01:46 crc kubenswrapper[4725]: I1202 14:01:46.435772 4725 generic.go:334] "Generic (PLEG): container finished" podID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerID="b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e" exitCode=0 Dec 02 14:01:46 crc kubenswrapper[4725]: I1202 14:01:46.435882 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnm28" event={"ID":"d8fd3136-ffdb-41b9-b49c-fce1777f4374","Type":"ContainerDied","Data":"b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e"} Dec 02 14:01:49 crc kubenswrapper[4725]: I1202 14:01:49.467280 4725 generic.go:334] "Generic (PLEG): container finished" podID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerID="8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555" exitCode=0 Dec 02 14:01:49 crc kubenswrapper[4725]: I1202 14:01:49.467486 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnm28" event={"ID":"d8fd3136-ffdb-41b9-b49c-fce1777f4374","Type":"ContainerDied","Data":"8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555"} Dec 02 14:01:55 crc kubenswrapper[4725]: I1202 14:01:55.517393 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnm28" event={"ID":"d8fd3136-ffdb-41b9-b49c-fce1777f4374","Type":"ContainerStarted","Data":"00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f"} Dec 02 14:01:56 crc kubenswrapper[4725]: I1202 14:01:56.551369 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wnm28" podStartSLOduration=6.987927627 podStartE2EDuration="12.551350085s" podCreationTimestamp="2025-12-02 14:01:44 +0000 UTC" firstStartedPulling="2025-12-02 14:01:46.43764642 +0000 UTC m=+3437.394288115" lastFinishedPulling="2025-12-02 14:01:52.001068878 +0000 UTC m=+3442.957710573" observedRunningTime="2025-12-02 14:01:56.543953611 +0000 UTC m=+3447.500595316" watchObservedRunningTime="2025-12-02 14:01:56.551350085 +0000 UTC m=+3447.507991780" Dec 02 14:02:04 crc kubenswrapper[4725]: I1202 14:02:04.805207 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:02:04 crc kubenswrapper[4725]: I1202 14:02:04.805856 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:02:04 crc kubenswrapper[4725]: I1202 14:02:04.905524 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:02:05 crc kubenswrapper[4725]: I1202 14:02:05.657618 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:02:05 crc kubenswrapper[4725]: I1202 14:02:05.706571 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wnm28"] Dec 02 14:02:07 crc kubenswrapper[4725]: I1202 14:02:07.628882 4725 generic.go:334] "Generic (PLEG): container finished" podID="ed2871c7-0ac2-42bf-9e37-3960588f4448" containerID="e652ccab5f70d4e41f7ce29e2c2f39d201dfcee921585830836390b17b685db4" exitCode=0 Dec 02 14:02:07 crc kubenswrapper[4725]: I1202 14:02:07.628972 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" event={"ID":"ed2871c7-0ac2-42bf-9e37-3960588f4448","Type":"ContainerDied","Data":"e652ccab5f70d4e41f7ce29e2c2f39d201dfcee921585830836390b17b685db4"} Dec 02 14:02:07 crc kubenswrapper[4725]: I1202 14:02:07.629495 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wnm28" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="registry-server" containerID="cri-o://00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f" gracePeriod=2 Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.076872 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.138189 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqln6\" (UniqueName: \"kubernetes.io/projected/d8fd3136-ffdb-41b9-b49c-fce1777f4374-kube-api-access-sqln6\") pod \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.148919 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8fd3136-ffdb-41b9-b49c-fce1777f4374-kube-api-access-sqln6" (OuterVolumeSpecName: "kube-api-access-sqln6") pod "d8fd3136-ffdb-41b9-b49c-fce1777f4374" (UID: "d8fd3136-ffdb-41b9-b49c-fce1777f4374"). InnerVolumeSpecName "kube-api-access-sqln6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.242558 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-utilities\") pod \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.242721 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-catalog-content\") pod \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\" (UID: \"d8fd3136-ffdb-41b9-b49c-fce1777f4374\") " Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.243548 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqln6\" (UniqueName: \"kubernetes.io/projected/d8fd3136-ffdb-41b9-b49c-fce1777f4374-kube-api-access-sqln6\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.243606 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-utilities" (OuterVolumeSpecName: "utilities") pod "d8fd3136-ffdb-41b9-b49c-fce1777f4374" (UID: "d8fd3136-ffdb-41b9-b49c-fce1777f4374"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.292658 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8fd3136-ffdb-41b9-b49c-fce1777f4374" (UID: "d8fd3136-ffdb-41b9-b49c-fce1777f4374"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.345014 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.345056 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8fd3136-ffdb-41b9-b49c-fce1777f4374-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.641073 4725 generic.go:334] "Generic (PLEG): container finished" podID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerID="00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f" exitCode=0 Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.641144 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnm28" event={"ID":"d8fd3136-ffdb-41b9-b49c-fce1777f4374","Type":"ContainerDied","Data":"00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f"} Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.641195 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wnm28" event={"ID":"d8fd3136-ffdb-41b9-b49c-fce1777f4374","Type":"ContainerDied","Data":"495bd1d671bb796f172e64bc5b9fb62af6d02acec0eb1f25a8801fc5524062e6"} Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.641213 4725 scope.go:117] "RemoveContainer" containerID="00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.646107 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wnm28" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.713646 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.732870 4725 scope.go:117] "RemoveContainer" containerID="8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.734292 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wnm28"] Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.749991 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wnm28"] Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.768607 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jn99z/crc-debug-v7gqc"] Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.771574 4725 scope.go:117] "RemoveContainer" containerID="b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.777642 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jn99z/crc-debug-v7gqc"] Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.830787 4725 scope.go:117] "RemoveContainer" containerID="00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f" Dec 02 14:02:08 crc kubenswrapper[4725]: E1202 14:02:08.831371 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f\": container with ID starting with 00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f not found: ID does not exist" containerID="00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.831522 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f"} err="failed to get container status \"00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f\": rpc error: code = NotFound desc = could not find container \"00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f\": container with ID starting with 00cbf75066f3e5959812e94c4dae127a87b6fd61df2e7936429e67188c38c56f not found: ID does not exist" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.831642 4725 scope.go:117] "RemoveContainer" containerID="8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555" Dec 02 14:02:08 crc kubenswrapper[4725]: E1202 14:02:08.832160 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555\": container with ID starting with 8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555 not found: ID does not exist" containerID="8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.832257 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555"} err="failed to get container status \"8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555\": rpc error: code = NotFound desc = could not find container \"8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555\": container with ID starting with 8e30bc28e66b40bc149e0e478aa67ccb64fc6189ecc99135fb3799db68fc2555 not found: ID does not exist" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.832345 4725 scope.go:117] "RemoveContainer" containerID="b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e" Dec 02 14:02:08 crc kubenswrapper[4725]: E1202 14:02:08.832674 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e\": container with ID starting with b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e not found: ID does not exist" containerID="b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.832778 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e"} err="failed to get container status \"b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e\": rpc error: code = NotFound desc = could not find container \"b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e\": container with ID starting with b0807f49a3e75987271b5df74ee7714d2f90f0dfd60cea2aee5813efed0bc19e not found: ID does not exist" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.852846 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed2871c7-0ac2-42bf-9e37-3960588f4448-host\") pod \"ed2871c7-0ac2-42bf-9e37-3960588f4448\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.853075 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2871c7-0ac2-42bf-9e37-3960588f4448-host" (OuterVolumeSpecName: "host") pod "ed2871c7-0ac2-42bf-9e37-3960588f4448" (UID: "ed2871c7-0ac2-42bf-9e37-3960588f4448"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.853249 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grxnl\" (UniqueName: \"kubernetes.io/projected/ed2871c7-0ac2-42bf-9e37-3960588f4448-kube-api-access-grxnl\") pod \"ed2871c7-0ac2-42bf-9e37-3960588f4448\" (UID: \"ed2871c7-0ac2-42bf-9e37-3960588f4448\") " Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.853854 4725 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ed2871c7-0ac2-42bf-9e37-3960588f4448-host\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.858744 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed2871c7-0ac2-42bf-9e37-3960588f4448-kube-api-access-grxnl" (OuterVolumeSpecName: "kube-api-access-grxnl") pod "ed2871c7-0ac2-42bf-9e37-3960588f4448" (UID: "ed2871c7-0ac2-42bf-9e37-3960588f4448"). InnerVolumeSpecName "kube-api-access-grxnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:02:08 crc kubenswrapper[4725]: I1202 14:02:08.956303 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grxnl\" (UniqueName: \"kubernetes.io/projected/ed2871c7-0ac2-42bf-9e37-3960588f4448-kube-api-access-grxnl\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.280520 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" path="/var/lib/kubelet/pods/d8fd3136-ffdb-41b9-b49c-fce1777f4374/volumes" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.281647 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed2871c7-0ac2-42bf-9e37-3960588f4448" path="/var/lib/kubelet/pods/ed2871c7-0ac2-42bf-9e37-3960588f4448/volumes" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.651397 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-v7gqc" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.651402 4725 scope.go:117] "RemoveContainer" containerID="e652ccab5f70d4e41f7ce29e2c2f39d201dfcee921585830836390b17b685db4" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.923638 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jn99z/crc-debug-zfnx5"] Dec 02 14:02:09 crc kubenswrapper[4725]: E1202 14:02:09.924134 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="registry-server" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.924158 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="registry-server" Dec 02 14:02:09 crc kubenswrapper[4725]: E1202 14:02:09.924193 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="extract-utilities" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.924201 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="extract-utilities" Dec 02 14:02:09 crc kubenswrapper[4725]: E1202 14:02:09.924216 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="extract-content" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.924227 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="extract-content" Dec 02 14:02:09 crc kubenswrapper[4725]: E1202 14:02:09.924246 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2871c7-0ac2-42bf-9e37-3960588f4448" containerName="container-00" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.924253 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2871c7-0ac2-42bf-9e37-3960588f4448" containerName="container-00" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.924506 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2871c7-0ac2-42bf-9e37-3960588f4448" containerName="container-00" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.924540 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8fd3136-ffdb-41b9-b49c-fce1777f4374" containerName="registry-server" Dec 02 14:02:09 crc kubenswrapper[4725]: I1202 14:02:09.925296 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.077767 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrfrh\" (UniqueName: \"kubernetes.io/projected/dcc14269-a303-4e64-8359-a6f9fe88ba2c-kube-api-access-qrfrh\") pod \"crc-debug-zfnx5\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.077918 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcc14269-a303-4e64-8359-a6f9fe88ba2c-host\") pod \"crc-debug-zfnx5\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.180124 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcc14269-a303-4e64-8359-a6f9fe88ba2c-host\") pod \"crc-debug-zfnx5\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.180297 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrfrh\" (UniqueName: \"kubernetes.io/projected/dcc14269-a303-4e64-8359-a6f9fe88ba2c-kube-api-access-qrfrh\") pod \"crc-debug-zfnx5\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.180306 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcc14269-a303-4e64-8359-a6f9fe88ba2c-host\") pod \"crc-debug-zfnx5\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.201593 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrfrh\" (UniqueName: \"kubernetes.io/projected/dcc14269-a303-4e64-8359-a6f9fe88ba2c-kube-api-access-qrfrh\") pod \"crc-debug-zfnx5\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.241380 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.662713 4725 generic.go:334] "Generic (PLEG): container finished" podID="dcc14269-a303-4e64-8359-a6f9fe88ba2c" containerID="746dbc3e0ea2e360ccf8c8c6ea56346cb22481d9b9055157f0ce3e5d282a9cab" exitCode=0 Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.662759 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-zfnx5" event={"ID":"dcc14269-a303-4e64-8359-a6f9fe88ba2c","Type":"ContainerDied","Data":"746dbc3e0ea2e360ccf8c8c6ea56346cb22481d9b9055157f0ce3e5d282a9cab"} Dec 02 14:02:10 crc kubenswrapper[4725]: I1202 14:02:10.662822 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-zfnx5" event={"ID":"dcc14269-a303-4e64-8359-a6f9fe88ba2c","Type":"ContainerStarted","Data":"a8a5dc7f668e8736c0d1da757ad6e18b8319f2f6e851cc7a73c39f0c1d550e4c"} Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.204509 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jn99z/crc-debug-zfnx5"] Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.219022 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jn99z/crc-debug-zfnx5"] Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.788341 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.918109 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrfrh\" (UniqueName: \"kubernetes.io/projected/dcc14269-a303-4e64-8359-a6f9fe88ba2c-kube-api-access-qrfrh\") pod \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.918311 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcc14269-a303-4e64-8359-a6f9fe88ba2c-host\") pod \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\" (UID: \"dcc14269-a303-4e64-8359-a6f9fe88ba2c\") " Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.918433 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc14269-a303-4e64-8359-a6f9fe88ba2c-host" (OuterVolumeSpecName: "host") pod "dcc14269-a303-4e64-8359-a6f9fe88ba2c" (UID: "dcc14269-a303-4e64-8359-a6f9fe88ba2c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.918961 4725 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcc14269-a303-4e64-8359-a6f9fe88ba2c-host\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:11 crc kubenswrapper[4725]: I1202 14:02:11.924033 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcc14269-a303-4e64-8359-a6f9fe88ba2c-kube-api-access-qrfrh" (OuterVolumeSpecName: "kube-api-access-qrfrh") pod "dcc14269-a303-4e64-8359-a6f9fe88ba2c" (UID: "dcc14269-a303-4e64-8359-a6f9fe88ba2c"). InnerVolumeSpecName "kube-api-access-qrfrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.020768 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrfrh\" (UniqueName: \"kubernetes.io/projected/dcc14269-a303-4e64-8359-a6f9fe88ba2c-kube-api-access-qrfrh\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.507324 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jn99z/crc-debug-ckktq"] Dec 02 14:02:12 crc kubenswrapper[4725]: E1202 14:02:12.507750 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc14269-a303-4e64-8359-a6f9fe88ba2c" containerName="container-00" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.507768 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc14269-a303-4e64-8359-a6f9fe88ba2c" containerName="container-00" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.507962 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc14269-a303-4e64-8359-a6f9fe88ba2c" containerName="container-00" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.509103 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.633350 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wq7s\" (UniqueName: \"kubernetes.io/projected/57f8ef9b-1892-455d-8565-03ea3d0812db-kube-api-access-8wq7s\") pod \"crc-debug-ckktq\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.633493 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57f8ef9b-1892-455d-8565-03ea3d0812db-host\") pod \"crc-debug-ckktq\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.680161 4725 scope.go:117] "RemoveContainer" containerID="746dbc3e0ea2e360ccf8c8c6ea56346cb22481d9b9055157f0ce3e5d282a9cab" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.680235 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-zfnx5" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.735182 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wq7s\" (UniqueName: \"kubernetes.io/projected/57f8ef9b-1892-455d-8565-03ea3d0812db-kube-api-access-8wq7s\") pod \"crc-debug-ckktq\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.735322 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57f8ef9b-1892-455d-8565-03ea3d0812db-host\") pod \"crc-debug-ckktq\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.735517 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57f8ef9b-1892-455d-8565-03ea3d0812db-host\") pod \"crc-debug-ckktq\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.757492 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wq7s\" (UniqueName: \"kubernetes.io/projected/57f8ef9b-1892-455d-8565-03ea3d0812db-kube-api-access-8wq7s\") pod \"crc-debug-ckktq\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: I1202 14:02:12.825892 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:12 crc kubenswrapper[4725]: W1202 14:02:12.852990 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57f8ef9b_1892_455d_8565_03ea3d0812db.slice/crio-6a7e9adb7a28e78654194c58089ef5937f88b861385f18f7a7cff326cf9f21fc WatchSource:0}: Error finding container 6a7e9adb7a28e78654194c58089ef5937f88b861385f18f7a7cff326cf9f21fc: Status 404 returned error can't find the container with id 6a7e9adb7a28e78654194c58089ef5937f88b861385f18f7a7cff326cf9f21fc Dec 02 14:02:13 crc kubenswrapper[4725]: I1202 14:02:13.278385 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcc14269-a303-4e64-8359-a6f9fe88ba2c" path="/var/lib/kubelet/pods/dcc14269-a303-4e64-8359-a6f9fe88ba2c/volumes" Dec 02 14:02:13 crc kubenswrapper[4725]: I1202 14:02:13.691095 4725 generic.go:334] "Generic (PLEG): container finished" podID="57f8ef9b-1892-455d-8565-03ea3d0812db" containerID="0f2a940ee3f141fd951531fbc902d4ce4885938aab37a1448aa76b064763f829" exitCode=0 Dec 02 14:02:13 crc kubenswrapper[4725]: I1202 14:02:13.691184 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-ckktq" event={"ID":"57f8ef9b-1892-455d-8565-03ea3d0812db","Type":"ContainerDied","Data":"0f2a940ee3f141fd951531fbc902d4ce4885938aab37a1448aa76b064763f829"} Dec 02 14:02:13 crc kubenswrapper[4725]: I1202 14:02:13.691232 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/crc-debug-ckktq" event={"ID":"57f8ef9b-1892-455d-8565-03ea3d0812db","Type":"ContainerStarted","Data":"6a7e9adb7a28e78654194c58089ef5937f88b861385f18f7a7cff326cf9f21fc"} Dec 02 14:02:13 crc kubenswrapper[4725]: I1202 14:02:13.742650 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jn99z/crc-debug-ckktq"] Dec 02 14:02:13 crc kubenswrapper[4725]: I1202 14:02:13.749538 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jn99z/crc-debug-ckktq"] Dec 02 14:02:14 crc kubenswrapper[4725]: I1202 14:02:14.817684 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:14 crc kubenswrapper[4725]: I1202 14:02:14.959116 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wq7s\" (UniqueName: \"kubernetes.io/projected/57f8ef9b-1892-455d-8565-03ea3d0812db-kube-api-access-8wq7s\") pod \"57f8ef9b-1892-455d-8565-03ea3d0812db\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " Dec 02 14:02:14 crc kubenswrapper[4725]: I1202 14:02:14.959422 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57f8ef9b-1892-455d-8565-03ea3d0812db-host\") pod \"57f8ef9b-1892-455d-8565-03ea3d0812db\" (UID: \"57f8ef9b-1892-455d-8565-03ea3d0812db\") " Dec 02 14:02:14 crc kubenswrapper[4725]: I1202 14:02:14.959554 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57f8ef9b-1892-455d-8565-03ea3d0812db-host" (OuterVolumeSpecName: "host") pod "57f8ef9b-1892-455d-8565-03ea3d0812db" (UID: "57f8ef9b-1892-455d-8565-03ea3d0812db"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 14:02:14 crc kubenswrapper[4725]: I1202 14:02:14.959987 4725 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/57f8ef9b-1892-455d-8565-03ea3d0812db-host\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:14 crc kubenswrapper[4725]: I1202 14:02:14.965398 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f8ef9b-1892-455d-8565-03ea3d0812db-kube-api-access-8wq7s" (OuterVolumeSpecName: "kube-api-access-8wq7s") pod "57f8ef9b-1892-455d-8565-03ea3d0812db" (UID: "57f8ef9b-1892-455d-8565-03ea3d0812db"). InnerVolumeSpecName "kube-api-access-8wq7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:02:15 crc kubenswrapper[4725]: I1202 14:02:15.062545 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wq7s\" (UniqueName: \"kubernetes.io/projected/57f8ef9b-1892-455d-8565-03ea3d0812db-kube-api-access-8wq7s\") on node \"crc\" DevicePath \"\"" Dec 02 14:02:15 crc kubenswrapper[4725]: I1202 14:02:15.287122 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f8ef9b-1892-455d-8565-03ea3d0812db" path="/var/lib/kubelet/pods/57f8ef9b-1892-455d-8565-03ea3d0812db/volumes" Dec 02 14:02:15 crc kubenswrapper[4725]: I1202 14:02:15.715363 4725 scope.go:117] "RemoveContainer" containerID="0f2a940ee3f141fd951531fbc902d4ce4885938aab37a1448aa76b064763f829" Dec 02 14:02:15 crc kubenswrapper[4725]: I1202 14:02:15.715406 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/crc-debug-ckktq" Dec 02 14:02:29 crc kubenswrapper[4725]: I1202 14:02:29.560367 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c6d675f88-kk42m_8063dc7f-156e-4f2f-a6fa-8759abb28c37/barbican-api/0.log" Dec 02 14:02:29 crc kubenswrapper[4725]: I1202 14:02:29.690821 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c6d675f88-kk42m_8063dc7f-156e-4f2f-a6fa-8759abb28c37/barbican-api-log/0.log" Dec 02 14:02:29 crc kubenswrapper[4725]: I1202 14:02:29.780676 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54b4d7cbbb-hdq62_e5f2ea09-3acf-4c71-8301-c6531e608bf1/barbican-keystone-listener/0.log" Dec 02 14:02:29 crc kubenswrapper[4725]: I1202 14:02:29.799735 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54b4d7cbbb-hdq62_e5f2ea09-3acf-4c71-8301-c6531e608bf1/barbican-keystone-listener-log/0.log" Dec 02 14:02:29 crc kubenswrapper[4725]: I1202 14:02:29.950883 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c5f8cd4f5-fd8q7_643b3918-6dab-4956-bd36-474ba0195d15/barbican-worker/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.017052 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c5f8cd4f5-fd8q7_643b3918-6dab-4956-bd36-474ba0195d15/barbican-worker-log/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.352177 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr_3ab59324-3446-4602-af02-e9094f220f34/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.393487 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/ceilometer-notification-agent/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.436492 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/ceilometer-central-agent/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.537904 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/proxy-httpd/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.606341 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/sg-core/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.686186 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eda48a5c-28d1-4b25-8377-a8870261551c/cinder-api/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.765908 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eda48a5c-28d1-4b25-8377-a8870261551c/cinder-api-log/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.871344 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3eab134b-db07-44ea-9f46-734dbda09036/cinder-scheduler/0.log" Dec 02 14:02:30 crc kubenswrapper[4725]: I1202 14:02:30.937765 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3eab134b-db07-44ea-9f46-734dbda09036/probe/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.091013 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc_711edc3c-a714-4e16-a7b8-6f61529f69d7/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.166483 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-d98hr_b013516f-cd7c-4a91-bc3e-bcddd68f439f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.302015 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-chnnr_cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61/init/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.499090 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-chnnr_cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61/init/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.518188 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-chnnr_cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61/dnsmasq-dns/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.573592 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg_022c7b77-3848-4ca6-9382-64eb9762bf40/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.741682 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2b150790-de06-4a61-901f-014c70552a9c/glance-httpd/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.774709 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2b150790-de06-4a61-901f-014c70552a9c/glance-log/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.968180 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17c38050-4a8a-45e9-8ff8-a40dffaee528/glance-log/0.log" Dec 02 14:02:31 crc kubenswrapper[4725]: I1202 14:02:31.971014 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17c38050-4a8a-45e9-8ff8-a40dffaee528/glance-httpd/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.236891 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f9f547644-22mp9_a4cfe8aa-4818-4626-8e86-6607e7bb38d2/horizon/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.303138 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8_5fee6c15-b89f-4d6b-8536-0e14c298e235/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.478817 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bxpzs_375b1a65-e012-4cfa-a08b-5d0b321be9d7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.509176 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f9f547644-22mp9_a4cfe8aa-4818-4626-8e86-6607e7bb38d2/horizon-log/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.797879 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29411401-76w49_7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70/keystone-cron/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.860063 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c7dc85fc9-tj9wl_16fb266f-4c1b-406e-b662-442a92366486/keystone-api/0.log" Dec 02 14:02:32 crc kubenswrapper[4725]: I1202 14:02:32.932670 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a093384f-2994-4207-8447-1cfe41c8a14e/kube-state-metrics/0.log" Dec 02 14:02:33 crc kubenswrapper[4725]: I1202 14:02:33.120837 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx_ee65f1b9-bbb7-4561-af64-e6f0ab3deec1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:33 crc kubenswrapper[4725]: I1202 14:02:33.440486 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b5b59474f-qnsqr_e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d/neutron-httpd/0.log" Dec 02 14:02:33 crc kubenswrapper[4725]: I1202 14:02:33.486164 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b5b59474f-qnsqr_e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d/neutron-api/0.log" Dec 02 14:02:33 crc kubenswrapper[4725]: I1202 14:02:33.890796 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk_d9633180-0a23-47b6-994f-13ec1eed9c30/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:34 crc kubenswrapper[4725]: I1202 14:02:34.364582 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_84138030-0cbd-41bd-8fb4-afe9fcf57bbc/nova-cell0-conductor-conductor/0.log" Dec 02 14:02:34 crc kubenswrapper[4725]: I1202 14:02:34.385572 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3aa1cff6-4533-4ba6-aeac-86a813cf8516/nova-api-log/0.log" Dec 02 14:02:34 crc kubenswrapper[4725]: I1202 14:02:34.527625 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3aa1cff6-4533-4ba6-aeac-86a813cf8516/nova-api-api/0.log" Dec 02 14:02:34 crc kubenswrapper[4725]: I1202 14:02:34.720035 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_71998d01-30ab-421a-bddc-5e49c6ac8b29/nova-cell1-conductor-conductor/0.log" Dec 02 14:02:34 crc kubenswrapper[4725]: I1202 14:02:34.731157 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_03aa0ff8-1156-466a-9726-48481ee110d8/nova-cell1-novncproxy-novncproxy/0.log" Dec 02 14:02:34 crc kubenswrapper[4725]: I1202 14:02:34.937005 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-58vqc_76bb9f49-7e28-49dc-9946-d3e6de9e6a26/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:35 crc kubenswrapper[4725]: I1202 14:02:35.079232 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d/nova-metadata-log/0.log" Dec 02 14:02:35 crc kubenswrapper[4725]: I1202 14:02:35.352286 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c1a212a8-1993-4689-988a-b6fc12ad4cb3/nova-scheduler-scheduler/0.log" Dec 02 14:02:35 crc kubenswrapper[4725]: I1202 14:02:35.439382 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc289673-6c5c-45f0-a9ff-df10cddd635c/mysql-bootstrap/0.log" Dec 02 14:02:35 crc kubenswrapper[4725]: I1202 14:02:35.591330 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc289673-6c5c-45f0-a9ff-df10cddd635c/mysql-bootstrap/0.log" Dec 02 14:02:35 crc kubenswrapper[4725]: I1202 14:02:35.672527 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc289673-6c5c-45f0-a9ff-df10cddd635c/galera/0.log" Dec 02 14:02:35 crc kubenswrapper[4725]: I1202 14:02:35.815178 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ef9cb921-2581-495e-be27-cae739b9a971/mysql-bootstrap/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.022712 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ef9cb921-2581-495e-be27-cae739b9a971/mysql-bootstrap/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.075852 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ef9cb921-2581-495e-be27-cae739b9a971/galera/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.141736 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d/nova-metadata-metadata/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.188888 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6ed64991-e56a-4635-b062-f68f7f1d134d/openstackclient/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.284686 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4ngrg_79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7/ovn-controller/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.416634 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kwn9x_4739bfa6-9e63-4f3a-bba3-b6c21afe2a74/openstack-network-exporter/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.566749 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovsdb-server-init/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.797042 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovsdb-server-init/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.812365 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovs-vswitchd/0.log" Dec 02 14:02:36 crc kubenswrapper[4725]: I1202 14:02:36.832213 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovsdb-server/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.051422 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_984ec540-6517-43f8-90e9-751b57ff0165/openstack-network-exporter/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.075559 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-rlrw6_9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.192321 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_984ec540-6517-43f8-90e9-751b57ff0165/ovn-northd/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.269324 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bad22ab8-0033-4a28-b7aa-65b1c285a989/openstack-network-exporter/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.456307 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bad22ab8-0033-4a28-b7aa-65b1c285a989/ovsdbserver-nb/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.553299 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_51144a39-c6dd-4a5f-adfb-6d62cb459485/openstack-network-exporter/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.590841 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_51144a39-c6dd-4a5f-adfb-6d62cb459485/ovsdbserver-sb/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.878264 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-784f98d458-wsm8f_66a69b36-610f-4f99-b57b-834d0a8a07a1/placement-api/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.887246 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1774e612-2433-4536-a07c-ef5ac77f2036/setup-container/0.log" Dec 02 14:02:37 crc kubenswrapper[4725]: I1202 14:02:37.906587 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-784f98d458-wsm8f_66a69b36-610f-4f99-b57b-834d0a8a07a1/placement-log/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.087521 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1774e612-2433-4536-a07c-ef5ac77f2036/setup-container/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.139221 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1774e612-2433-4536-a07c-ef5ac77f2036/rabbitmq/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.199776 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7/setup-container/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.431342 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7/setup-container/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.462347 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7/rabbitmq/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.525421 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd_0372a2f9-72d1-468a-bd81-5a1a0a48c545/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.710659 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-mjvdg_efdcb239-8753-47be-86b7-267371f72e53/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.811331 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h_8994bc56-f6d8-49d0-8236-eab3fe2d35e7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:38 crc kubenswrapper[4725]: I1202 14:02:38.936000 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-twz8w_2c170bed-2666-437d-b8dc-b14bb81bd6eb/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.044017 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9shf4_50bb29f4-5ab8-421c-a5f5-4321d106e1fe/ssh-known-hosts-edpm-deployment/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.253160 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d448f6767-ccmmh_69be1bb2-b1e2-41fb-991c-c6a12db18fb8/proxy-httpd/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.279842 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d448f6767-ccmmh_69be1bb2-b1e2-41fb-991c-c6a12db18fb8/proxy-server/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.407300 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-z72np_dda4b826-1871-43b5-808d-45d84f2f2483/swift-ring-rebalance/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.511576 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-auditor/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.631231 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-reaper/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.663982 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-replicator/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.732747 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-auditor/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.782134 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-server/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.846997 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-server/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.877595 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-replicator/0.log" Dec 02 14:02:39 crc kubenswrapper[4725]: I1202 14:02:39.954499 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-updater/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.038200 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-auditor/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.068640 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-expirer/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.142192 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-replicator/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.153182 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-server/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.294749 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-updater/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.360322 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/swift-recon-cron/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.363421 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/rsync/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.559875 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc_b45407f7-eb82-40f8-a223-0dfbabc966e5/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.629117 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8/tempest-tests-tempest-tests-runner/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.844452 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_32b05d70-2f30-48dd-8fba-8187969352da/test-operator-logs-container/0.log" Dec 02 14:02:40 crc kubenswrapper[4725]: I1202 14:02:40.892527 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-dgwld_4a420694-62a2-4fd1-aade-ebfaaf73590c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:02:47 crc kubenswrapper[4725]: I1202 14:02:47.801046 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_21bbf393-00ed-46c6-a61f-8a458212e8e7/memcached/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.545606 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/util/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.740401 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/util/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.753033 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/pull/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.754326 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/pull/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.899229 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/util/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.922409 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/pull/0.log" Dec 02 14:03:05 crc kubenswrapper[4725]: I1202 14:03:05.928828 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/extract/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.099905 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ctzjq_b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b/kube-rbac-proxy/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.142877 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-z9ttg_570a3a73-dffb-4ded-8fd1-d30ddc60979d/kube-rbac-proxy/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.152233 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ctzjq_b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b/manager/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.314323 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-ppgnw_c42118d3-322d-4733-8749-d13bc174a6f8/kube-rbac-proxy/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.331121 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-z9ttg_570a3a73-dffb-4ded-8fd1-d30ddc60979d/manager/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.383375 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-ppgnw_c42118d3-322d-4733-8749-d13bc174a6f8/manager/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.531357 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-njlj7_65a9f532-387a-487e-bfe0-81a110509366/kube-rbac-proxy/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.648816 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-njlj7_65a9f532-387a-487e-bfe0-81a110509366/manager/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.708656 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cwfk4_0b562f2a-162c-4611-81d3-00937e945f96/kube-rbac-proxy/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.733436 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cwfk4_0b562f2a-162c-4611-81d3-00937e945f96/manager/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.857030 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pmdw6_6b08a8fe-08d5-4b04-9283-51973a9ac3d6/kube-rbac-proxy/0.log" Dec 02 14:03:06 crc kubenswrapper[4725]: I1202 14:03:06.919746 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pmdw6_6b08a8fe-08d5-4b04-9283-51973a9ac3d6/manager/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.015225 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g99db_bf404326-3c43-4cc4-93b7-0793213afcba/kube-rbac-proxy/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.220133 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-klrxl_4bad83be-e3ed-499a-8e64-cfe83c4ac1e6/kube-rbac-proxy/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.240000 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-klrxl_4bad83be-e3ed-499a-8e64-cfe83c4ac1e6/manager/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.242417 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g99db_bf404326-3c43-4cc4-93b7-0793213afcba/manager/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.399365 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p22fj_3fb7e977-cd14-4b71-b349-19d4487cfb15/kube-rbac-proxy/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.560741 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p22fj_3fb7e977-cd14-4b71-b349-19d4487cfb15/manager/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.573214 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rk99q_c1f336f6-716b-4096-89cd-f23d1fed90cb/kube-rbac-proxy/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.647763 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rk99q_c1f336f6-716b-4096-89cd-f23d1fed90cb/manager/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.750012 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-8s44s_8f32e184-607a-4077-aa41-ca7aabba76aa/kube-rbac-proxy/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.758619 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-8s44s_8f32e184-607a-4077-aa41-ca7aabba76aa/manager/0.log" Dec 02 14:03:07 crc kubenswrapper[4725]: I1202 14:03:07.936583 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-wd49v_2c988ae2-9613-4584-a599-a7d063611189/kube-rbac-proxy/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.022413 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-wd49v_2c988ae2-9613-4584-a599-a7d063611189/manager/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.175161 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-cllpp_eacfb091-583b-4e89-9850-cc146cef7e20/kube-rbac-proxy/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.254760 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-cllpp_eacfb091-583b-4e89-9850-cc146cef7e20/manager/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.306081 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-zdj64_bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b/kube-rbac-proxy/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.355227 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-zdj64_bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b/manager/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.436248 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc_9c4065a8-50f2-4be9-bdd7-42345814f4c5/kube-rbac-proxy/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.504267 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc_9c4065a8-50f2-4be9-bdd7-42345814f4c5/manager/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.887864 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-59k9r_6f802fa6-11b4-40ef-b9f1-bfb654645474/registry-server/0.log" Dec 02 14:03:08 crc kubenswrapper[4725]: I1202 14:03:08.926137 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75ff9f7b7-wcxkf_98fc37a5-6d29-4f0b-93ab-ffecc2157b33/operator/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.065992 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mxpqd_511bab3c-3885-4c5a-8c91-bd231b1c2a03/kube-rbac-proxy/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.238017 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mxpqd_511bab3c-3885-4c5a-8c91-bd231b1c2a03/manager/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.282844 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wqn2j_ad9c89d4-9792-401b-9f1f-81d2e5c7ede9/kube-rbac-proxy/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.463296 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wqn2j_ad9c89d4-9792-401b-9f1f-81d2e5c7ede9/manager/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.597869 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-ctlsq_6b03a06a-b4d5-4443-98eb-fdea3fd8302c/operator/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.669577 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-cptsj_dc849752-f58a-4e7a-b1fa-5b7d88c14d1a/kube-rbac-proxy/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.743396 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-cptsj_dc849752-f58a-4e7a-b1fa-5b7d88c14d1a/manager/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.837009 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-2l2x4_bd74c0bd-9ef5-487c-b190-2b23a8120c8d/kube-rbac-proxy/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.847713 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-754459f956-vkwfn_e68bdc86-d09e-48ae-839b-9866e70e1383/manager/0.log" Dec 02 14:03:09 crc kubenswrapper[4725]: I1202 14:03:09.947951 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-2l2x4_bd74c0bd-9ef5-487c-b190-2b23a8120c8d/manager/0.log" Dec 02 14:03:10 crc kubenswrapper[4725]: I1202 14:03:10.024240 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-nrrpt_7965a5b4-89be-44c6-8471-20ede0f6c92f/kube-rbac-proxy/0.log" Dec 02 14:03:10 crc kubenswrapper[4725]: I1202 14:03:10.032386 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-nrrpt_7965a5b4-89be-44c6-8471-20ede0f6c92f/manager/0.log" Dec 02 14:03:10 crc kubenswrapper[4725]: I1202 14:03:10.148903 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-rrnmn_b71451df-d329-44b4-8d63-f0ea35db6ce7/kube-rbac-proxy/0.log" Dec 02 14:03:10 crc kubenswrapper[4725]: I1202 14:03:10.191819 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-rrnmn_b71451df-d329-44b4-8d63-f0ea35db6ce7/manager/0.log" Dec 02 14:03:27 crc kubenswrapper[4725]: I1202 14:03:27.497355 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bphpt_4d37326e-fdd1-41e0-8401-5f47038356b3/control-plane-machine-set-operator/0.log" Dec 02 14:03:27 crc kubenswrapper[4725]: I1202 14:03:27.684325 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h8fhj_c17fbfab-7855-44b3-a983-22915062793a/kube-rbac-proxy/0.log" Dec 02 14:03:27 crc kubenswrapper[4725]: I1202 14:03:27.717182 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h8fhj_c17fbfab-7855-44b3-a983-22915062793a/machine-api-operator/0.log" Dec 02 14:03:39 crc kubenswrapper[4725]: I1202 14:03:39.361935 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6prlb_81141996-2325-4372-81be-7f69cbec7194/cert-manager-controller/0.log" Dec 02 14:03:39 crc kubenswrapper[4725]: I1202 14:03:39.482953 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-f864q_864d4bd0-504a-4d91-bb28-b920fe8845aa/cert-manager-cainjector/0.log" Dec 02 14:03:39 crc kubenswrapper[4725]: I1202 14:03:39.552342 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hsrpm_f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4/cert-manager-webhook/0.log" Dec 02 14:03:50 crc kubenswrapper[4725]: I1202 14:03:50.688976 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5dmqd_a4844270-686d-4f0b-aafc-5b2b1924444f/nmstate-console-plugin/0.log" Dec 02 14:03:50 crc kubenswrapper[4725]: I1202 14:03:50.855929 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6cqs8_28a8d066-b52d-450e-a738-884b8b260da5/nmstate-handler/0.log" Dec 02 14:03:50 crc kubenswrapper[4725]: I1202 14:03:50.902692 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-v7ngp_93447f0a-ea95-4ec5-beea-8733e261d5f3/kube-rbac-proxy/0.log" Dec 02 14:03:50 crc kubenswrapper[4725]: I1202 14:03:50.969604 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-v7ngp_93447f0a-ea95-4ec5-beea-8733e261d5f3/nmstate-metrics/0.log" Dec 02 14:03:51 crc kubenswrapper[4725]: I1202 14:03:51.041857 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-xz2lq_511d65e3-135b-4218-aaeb-bee781e9befb/nmstate-operator/0.log" Dec 02 14:03:51 crc kubenswrapper[4725]: I1202 14:03:51.160817 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-9rkbp_15ebd594-c399-41a5-a80a-6ce365f9e657/nmstate-webhook/0.log" Dec 02 14:03:54 crc kubenswrapper[4725]: I1202 14:03:54.284938 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:03:54 crc kubenswrapper[4725]: I1202 14:03:54.287582 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.489326 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bkrzn_ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b/kube-rbac-proxy/0.log" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.586979 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bkrzn_ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b/controller/0.log" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.707067 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.877738 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.907945 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.930491 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:04:04 crc kubenswrapper[4725]: I1202 14:04:04.959045 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.128528 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.136013 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.137690 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.157871 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.319792 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.359718 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.380133 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/controller/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.385108 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.506040 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/frr-metrics/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.586413 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/kube-rbac-proxy-frr/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.587093 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/kube-rbac-proxy/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.804013 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c7g46_00168d15-ce5e-4164-891c-1403031bee1d/frr-k8s-webhook-server/0.log" Dec 02 14:04:05 crc kubenswrapper[4725]: I1202 14:04:05.843068 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/reloader/0.log" Dec 02 14:04:06 crc kubenswrapper[4725]: I1202 14:04:06.077385 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-747844468f-qq5ns_bc5b46e6-ecea-4260-a386-e7408c29a842/manager/0.log" Dec 02 14:04:06 crc kubenswrapper[4725]: I1202 14:04:06.259862 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-857fd986b8-tqd7q_2315c4fc-8de9-4d85-9fab-a2cedd0eb95d/webhook-server/0.log" Dec 02 14:04:06 crc kubenswrapper[4725]: I1202 14:04:06.378070 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dhmrd_ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb/kube-rbac-proxy/0.log" Dec 02 14:04:06 crc kubenswrapper[4725]: I1202 14:04:06.933151 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dhmrd_ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb/speaker/0.log" Dec 02 14:04:06 crc kubenswrapper[4725]: I1202 14:04:06.975038 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/frr/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.378368 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/util/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.566859 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/util/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.582138 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/pull/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.610192 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/pull/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.749396 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/util/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.763112 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/pull/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.767946 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/extract/0.log" Dec 02 14:04:18 crc kubenswrapper[4725]: I1202 14:04:18.956576 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/util/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.087241 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/util/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.096379 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/pull/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.103340 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/pull/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.274255 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/util/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.275156 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/pull/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.277368 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/extract/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.449857 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-utilities/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.599131 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-utilities/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.660273 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-content/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.660517 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-content/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.859628 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-content/0.log" Dec 02 14:04:19 crc kubenswrapper[4725]: I1202 14:04:19.878596 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-utilities/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.070927 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-utilities/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.238966 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-utilities/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.299850 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-content/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.305027 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-content/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.432392 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/registry-server/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.479893 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-utilities/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.491021 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-content/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.669558 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kcshx_55b1c898-ab77-4a63-ba54-4c3246ec5732/marketplace-operator/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.894710 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-utilities/0.log" Dec 02 14:04:20 crc kubenswrapper[4725]: I1202 14:04:20.991495 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/registry-server/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.091777 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-content/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.110636 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-content/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.115736 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-utilities/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.271803 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-content/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.306772 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-utilities/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.438059 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/registry-server/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.505990 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-utilities/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.633669 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-utilities/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.665860 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-content/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.692444 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-content/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.856168 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-utilities/0.log" Dec 02 14:04:21 crc kubenswrapper[4725]: I1202 14:04:21.877469 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-content/0.log" Dec 02 14:04:22 crc kubenswrapper[4725]: I1202 14:04:22.241328 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/registry-server/0.log" Dec 02 14:04:24 crc kubenswrapper[4725]: I1202 14:04:24.285081 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:04:24 crc kubenswrapper[4725]: I1202 14:04:24.285391 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:04:54 crc kubenswrapper[4725]: I1202 14:04:54.284677 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:04:54 crc kubenswrapper[4725]: I1202 14:04:54.285351 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:04:54 crc kubenswrapper[4725]: I1202 14:04:54.285432 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 14:04:54 crc kubenswrapper[4725]: I1202 14:04:54.286402 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 14:04:54 crc kubenswrapper[4725]: I1202 14:04:54.286557 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" gracePeriod=600 Dec 02 14:04:54 crc kubenswrapper[4725]: E1202 14:04:54.423600 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:04:55 crc kubenswrapper[4725]: I1202 14:04:55.138888 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" exitCode=0 Dec 02 14:04:55 crc kubenswrapper[4725]: I1202 14:04:55.138931 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376"} Dec 02 14:04:55 crc kubenswrapper[4725]: I1202 14:04:55.138973 4725 scope.go:117] "RemoveContainer" containerID="cadf99b0b053173fd0df960b6f5ab3a95c2f07c85dde7cc800424b5829a75d8b" Dec 02 14:04:55 crc kubenswrapper[4725]: I1202 14:04:55.140557 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:04:55 crc kubenswrapper[4725]: E1202 14:04:55.141326 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:05:07 crc kubenswrapper[4725]: I1202 14:05:07.269453 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:05:07 crc kubenswrapper[4725]: E1202 14:05:07.270615 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:05:22 crc kubenswrapper[4725]: I1202 14:05:22.267876 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:05:22 crc kubenswrapper[4725]: E1202 14:05:22.268678 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:05:35 crc kubenswrapper[4725]: I1202 14:05:35.269099 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:05:35 crc kubenswrapper[4725]: E1202 14:05:35.269858 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.363796 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k7tv2"] Dec 02 14:05:38 crc kubenswrapper[4725]: E1202 14:05:38.364727 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f8ef9b-1892-455d-8565-03ea3d0812db" containerName="container-00" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.364741 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f8ef9b-1892-455d-8565-03ea3d0812db" containerName="container-00" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.365054 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f8ef9b-1892-455d-8565-03ea3d0812db" containerName="container-00" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.368417 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.391484 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k7tv2"] Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.503396 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5qm\" (UniqueName: \"kubernetes.io/projected/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-kube-api-access-tl5qm\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.503912 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-catalog-content\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.504145 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-utilities\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.606695 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5qm\" (UniqueName: \"kubernetes.io/projected/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-kube-api-access-tl5qm\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.607103 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-catalog-content\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.607245 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-utilities\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.608028 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-utilities\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.608698 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-catalog-content\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.633030 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5qm\" (UniqueName: \"kubernetes.io/projected/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-kube-api-access-tl5qm\") pod \"community-operators-k7tv2\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:38 crc kubenswrapper[4725]: I1202 14:05:38.700614 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:39 crc kubenswrapper[4725]: I1202 14:05:39.210928 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k7tv2"] Dec 02 14:05:39 crc kubenswrapper[4725]: I1202 14:05:39.555005 4725 generic.go:334] "Generic (PLEG): container finished" podID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerID="ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635" exitCode=0 Dec 02 14:05:39 crc kubenswrapper[4725]: I1202 14:05:39.555046 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7tv2" event={"ID":"f66f4e11-e42a-4c08-a21a-4d3ea808a89a","Type":"ContainerDied","Data":"ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635"} Dec 02 14:05:39 crc kubenswrapper[4725]: I1202 14:05:39.555246 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7tv2" event={"ID":"f66f4e11-e42a-4c08-a21a-4d3ea808a89a","Type":"ContainerStarted","Data":"d7f665da03dc9f25186b04a31a933784313e90023990e1c38a3ec621cf8ef22c"} Dec 02 14:05:39 crc kubenswrapper[4725]: I1202 14:05:39.556973 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.549066 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l567r"] Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.551726 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.570662 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l567r"] Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.651574 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9p42\" (UniqueName: \"kubernetes.io/projected/11ac3946-a0f4-431e-965b-621442994ad6-kube-api-access-r9p42\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.651622 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-utilities\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.651938 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-catalog-content\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.753953 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-catalog-content\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.754081 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9p42\" (UniqueName: \"kubernetes.io/projected/11ac3946-a0f4-431e-965b-621442994ad6-kube-api-access-r9p42\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.754109 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-utilities\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.754508 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-catalog-content\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.754521 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-utilities\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.777131 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9p42\" (UniqueName: \"kubernetes.io/projected/11ac3946-a0f4-431e-965b-621442994ad6-kube-api-access-r9p42\") pod \"redhat-marketplace-l567r\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:40 crc kubenswrapper[4725]: I1202 14:05:40.881018 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:41 crc kubenswrapper[4725]: I1202 14:05:41.346440 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l567r"] Dec 02 14:05:41 crc kubenswrapper[4725]: I1202 14:05:41.576120 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerStarted","Data":"a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5"} Dec 02 14:05:41 crc kubenswrapper[4725]: I1202 14:05:41.576410 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerStarted","Data":"604f99a474dca46e733e5a389a980fe5abee038aed328f80f7730ee09a689343"} Dec 02 14:05:42 crc kubenswrapper[4725]: I1202 14:05:42.589656 4725 generic.go:334] "Generic (PLEG): container finished" podID="11ac3946-a0f4-431e-965b-621442994ad6" containerID="a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5" exitCode=0 Dec 02 14:05:42 crc kubenswrapper[4725]: I1202 14:05:42.589928 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerDied","Data":"a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5"} Dec 02 14:05:42 crc kubenswrapper[4725]: I1202 14:05:42.594922 4725 generic.go:334] "Generic (PLEG): container finished" podID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerID="fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e" exitCode=0 Dec 02 14:05:42 crc kubenswrapper[4725]: I1202 14:05:42.594968 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7tv2" event={"ID":"f66f4e11-e42a-4c08-a21a-4d3ea808a89a","Type":"ContainerDied","Data":"fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e"} Dec 02 14:05:43 crc kubenswrapper[4725]: I1202 14:05:43.627357 4725 generic.go:334] "Generic (PLEG): container finished" podID="11ac3946-a0f4-431e-965b-621442994ad6" containerID="6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505" exitCode=0 Dec 02 14:05:43 crc kubenswrapper[4725]: I1202 14:05:43.628444 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerDied","Data":"6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505"} Dec 02 14:05:44 crc kubenswrapper[4725]: I1202 14:05:44.642776 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7tv2" event={"ID":"f66f4e11-e42a-4c08-a21a-4d3ea808a89a","Type":"ContainerStarted","Data":"615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7"} Dec 02 14:05:44 crc kubenswrapper[4725]: I1202 14:05:44.673653 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k7tv2" podStartSLOduration=2.67763136 podStartE2EDuration="6.673629868s" podCreationTimestamp="2025-12-02 14:05:38 +0000 UTC" firstStartedPulling="2025-12-02 14:05:39.556688011 +0000 UTC m=+3670.513329706" lastFinishedPulling="2025-12-02 14:05:43.552686519 +0000 UTC m=+3674.509328214" observedRunningTime="2025-12-02 14:05:44.663749983 +0000 UTC m=+3675.620391678" watchObservedRunningTime="2025-12-02 14:05:44.673629868 +0000 UTC m=+3675.630271563" Dec 02 14:05:45 crc kubenswrapper[4725]: I1202 14:05:45.659475 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerStarted","Data":"0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2"} Dec 02 14:05:45 crc kubenswrapper[4725]: I1202 14:05:45.687279 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l567r" podStartSLOduration=3.825519962 podStartE2EDuration="5.687259409s" podCreationTimestamp="2025-12-02 14:05:40 +0000 UTC" firstStartedPulling="2025-12-02 14:05:42.592697172 +0000 UTC m=+3673.549338867" lastFinishedPulling="2025-12-02 14:05:44.454436619 +0000 UTC m=+3675.411078314" observedRunningTime="2025-12-02 14:05:45.681201468 +0000 UTC m=+3676.637843173" watchObservedRunningTime="2025-12-02 14:05:45.687259409 +0000 UTC m=+3676.643901104" Dec 02 14:05:48 crc kubenswrapper[4725]: I1202 14:05:48.701503 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:48 crc kubenswrapper[4725]: I1202 14:05:48.703083 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:48 crc kubenswrapper[4725]: I1202 14:05:48.747812 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:49 crc kubenswrapper[4725]: I1202 14:05:49.278671 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:05:49 crc kubenswrapper[4725]: E1202 14:05:49.279577 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:05:49 crc kubenswrapper[4725]: I1202 14:05:49.742373 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:50 crc kubenswrapper[4725]: I1202 14:05:50.154709 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k7tv2"] Dec 02 14:05:50 crc kubenswrapper[4725]: I1202 14:05:50.881276 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:50 crc kubenswrapper[4725]: I1202 14:05:50.881336 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:50 crc kubenswrapper[4725]: I1202 14:05:50.928973 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:51 crc kubenswrapper[4725]: I1202 14:05:51.709187 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k7tv2" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="registry-server" containerID="cri-o://615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7" gracePeriod=2 Dec 02 14:05:51 crc kubenswrapper[4725]: I1202 14:05:51.763005 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.465036 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.535784 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l567r"] Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.614275 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl5qm\" (UniqueName: \"kubernetes.io/projected/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-kube-api-access-tl5qm\") pod \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.614643 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-utilities\") pod \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.614705 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-catalog-content\") pod \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\" (UID: \"f66f4e11-e42a-4c08-a21a-4d3ea808a89a\") " Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.615670 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-utilities" (OuterVolumeSpecName: "utilities") pod "f66f4e11-e42a-4c08-a21a-4d3ea808a89a" (UID: "f66f4e11-e42a-4c08-a21a-4d3ea808a89a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.621528 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-kube-api-access-tl5qm" (OuterVolumeSpecName: "kube-api-access-tl5qm") pod "f66f4e11-e42a-4c08-a21a-4d3ea808a89a" (UID: "f66f4e11-e42a-4c08-a21a-4d3ea808a89a"). InnerVolumeSpecName "kube-api-access-tl5qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.717598 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl5qm\" (UniqueName: \"kubernetes.io/projected/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-kube-api-access-tl5qm\") on node \"crc\" DevicePath \"\"" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.717639 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.736710 4725 generic.go:334] "Generic (PLEG): container finished" podID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerID="615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7" exitCode=0 Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.736863 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k7tv2" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.736786 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7tv2" event={"ID":"f66f4e11-e42a-4c08-a21a-4d3ea808a89a","Type":"ContainerDied","Data":"615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7"} Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.737099 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k7tv2" event={"ID":"f66f4e11-e42a-4c08-a21a-4d3ea808a89a","Type":"ContainerDied","Data":"d7f665da03dc9f25186b04a31a933784313e90023990e1c38a3ec621cf8ef22c"} Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.737150 4725 scope.go:117] "RemoveContainer" containerID="615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.749999 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f66f4e11-e42a-4c08-a21a-4d3ea808a89a" (UID: "f66f4e11-e42a-4c08-a21a-4d3ea808a89a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.766190 4725 scope.go:117] "RemoveContainer" containerID="fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.792335 4725 scope.go:117] "RemoveContainer" containerID="ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.819890 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f66f4e11-e42a-4c08-a21a-4d3ea808a89a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.832876 4725 scope.go:117] "RemoveContainer" containerID="615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7" Dec 02 14:05:52 crc kubenswrapper[4725]: E1202 14:05:52.833523 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7\": container with ID starting with 615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7 not found: ID does not exist" containerID="615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.833569 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7"} err="failed to get container status \"615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7\": rpc error: code = NotFound desc = could not find container \"615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7\": container with ID starting with 615d73efc47c7c2ae1c095853609d735b72a27931024bee8e07b80c722b8fea7 not found: ID does not exist" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.833598 4725 scope.go:117] "RemoveContainer" containerID="fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e" Dec 02 14:05:52 crc kubenswrapper[4725]: E1202 14:05:52.833965 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e\": container with ID starting with fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e not found: ID does not exist" containerID="fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.834001 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e"} err="failed to get container status \"fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e\": rpc error: code = NotFound desc = could not find container \"fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e\": container with ID starting with fcd70927d3049672bea8003f0d8e0878d65711b1d6c0337c3d112d1cf614e19e not found: ID does not exist" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.834026 4725 scope.go:117] "RemoveContainer" containerID="ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635" Dec 02 14:05:52 crc kubenswrapper[4725]: E1202 14:05:52.834303 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635\": container with ID starting with ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635 not found: ID does not exist" containerID="ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635" Dec 02 14:05:52 crc kubenswrapper[4725]: I1202 14:05:52.834335 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635"} err="failed to get container status \"ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635\": rpc error: code = NotFound desc = could not find container \"ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635\": container with ID starting with ab45abd09215b07c321727a8ce6f15f0cd9436f34d026b9a0ef6ce2bb3a96635 not found: ID does not exist" Dec 02 14:05:53 crc kubenswrapper[4725]: I1202 14:05:53.114550 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k7tv2"] Dec 02 14:05:53 crc kubenswrapper[4725]: I1202 14:05:53.122626 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k7tv2"] Dec 02 14:05:53 crc kubenswrapper[4725]: I1202 14:05:53.282637 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" path="/var/lib/kubelet/pods/f66f4e11-e42a-4c08-a21a-4d3ea808a89a/volumes" Dec 02 14:05:53 crc kubenswrapper[4725]: I1202 14:05:53.751496 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l567r" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="registry-server" containerID="cri-o://0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2" gracePeriod=2 Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.230313 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.372032 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-catalog-content\") pod \"11ac3946-a0f4-431e-965b-621442994ad6\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.372209 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9p42\" (UniqueName: \"kubernetes.io/projected/11ac3946-a0f4-431e-965b-621442994ad6-kube-api-access-r9p42\") pod \"11ac3946-a0f4-431e-965b-621442994ad6\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.372266 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-utilities\") pod \"11ac3946-a0f4-431e-965b-621442994ad6\" (UID: \"11ac3946-a0f4-431e-965b-621442994ad6\") " Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.373499 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-utilities" (OuterVolumeSpecName: "utilities") pod "11ac3946-a0f4-431e-965b-621442994ad6" (UID: "11ac3946-a0f4-431e-965b-621442994ad6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.377442 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ac3946-a0f4-431e-965b-621442994ad6-kube-api-access-r9p42" (OuterVolumeSpecName: "kube-api-access-r9p42") pod "11ac3946-a0f4-431e-965b-621442994ad6" (UID: "11ac3946-a0f4-431e-965b-621442994ad6"). InnerVolumeSpecName "kube-api-access-r9p42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.406400 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11ac3946-a0f4-431e-965b-621442994ad6" (UID: "11ac3946-a0f4-431e-965b-621442994ad6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.476929 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.476965 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9p42\" (UniqueName: \"kubernetes.io/projected/11ac3946-a0f4-431e-965b-621442994ad6-kube-api-access-r9p42\") on node \"crc\" DevicePath \"\"" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.476987 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11ac3946-a0f4-431e-965b-621442994ad6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.763688 4725 generic.go:334] "Generic (PLEG): container finished" podID="11ac3946-a0f4-431e-965b-621442994ad6" containerID="0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2" exitCode=0 Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.763742 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerDied","Data":"0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2"} Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.764040 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l567r" event={"ID":"11ac3946-a0f4-431e-965b-621442994ad6","Type":"ContainerDied","Data":"604f99a474dca46e733e5a389a980fe5abee038aed328f80f7730ee09a689343"} Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.764066 4725 scope.go:117] "RemoveContainer" containerID="0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.763849 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l567r" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.791896 4725 scope.go:117] "RemoveContainer" containerID="6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.818834 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l567r"] Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.828879 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l567r"] Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.830813 4725 scope.go:117] "RemoveContainer" containerID="a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.868236 4725 scope.go:117] "RemoveContainer" containerID="0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2" Dec 02 14:05:54 crc kubenswrapper[4725]: E1202 14:05:54.868805 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2\": container with ID starting with 0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2 not found: ID does not exist" containerID="0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.868877 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2"} err="failed to get container status \"0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2\": rpc error: code = NotFound desc = could not find container \"0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2\": container with ID starting with 0b78ce1c2dd4be2cfd331821bea3f6fac040509a6ac8f1b971203683a02e29c2 not found: ID does not exist" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.868913 4725 scope.go:117] "RemoveContainer" containerID="6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505" Dec 02 14:05:54 crc kubenswrapper[4725]: E1202 14:05:54.870139 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505\": container with ID starting with 6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505 not found: ID does not exist" containerID="6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.870185 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505"} err="failed to get container status \"6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505\": rpc error: code = NotFound desc = could not find container \"6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505\": container with ID starting with 6bf04d2c703e74cf7831355acf84ba3e8ad96ab3b8bbd0de55b398b842a89505 not found: ID does not exist" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.870209 4725 scope.go:117] "RemoveContainer" containerID="a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5" Dec 02 14:05:54 crc kubenswrapper[4725]: E1202 14:05:54.870654 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5\": container with ID starting with a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5 not found: ID does not exist" containerID="a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5" Dec 02 14:05:54 crc kubenswrapper[4725]: I1202 14:05:54.870688 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5"} err="failed to get container status \"a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5\": rpc error: code = NotFound desc = could not find container \"a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5\": container with ID starting with a67c4142dbe4f38bd4cfe421aa39cd5433cef08629f4a90a078212cb96816ca5 not found: ID does not exist" Dec 02 14:05:55 crc kubenswrapper[4725]: I1202 14:05:55.278950 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ac3946-a0f4-431e-965b-621442994ad6" path="/var/lib/kubelet/pods/11ac3946-a0f4-431e-965b-621442994ad6/volumes" Dec 02 14:06:01 crc kubenswrapper[4725]: I1202 14:06:01.863554 4725 generic.go:334] "Generic (PLEG): container finished" podID="54433b1a-0157-4690-bffb-01bb356b64fc" containerID="90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7" exitCode=0 Dec 02 14:06:01 crc kubenswrapper[4725]: I1202 14:06:01.863643 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jn99z/must-gather-dpsd4" event={"ID":"54433b1a-0157-4690-bffb-01bb356b64fc","Type":"ContainerDied","Data":"90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7"} Dec 02 14:06:01 crc kubenswrapper[4725]: I1202 14:06:01.864762 4725 scope.go:117] "RemoveContainer" containerID="90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7" Dec 02 14:06:02 crc kubenswrapper[4725]: I1202 14:06:02.105297 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jn99z_must-gather-dpsd4_54433b1a-0157-4690-bffb-01bb356b64fc/gather/0.log" Dec 02 14:06:02 crc kubenswrapper[4725]: I1202 14:06:02.269775 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:06:02 crc kubenswrapper[4725]: E1202 14:06:02.270431 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.432446 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jn99z/must-gather-dpsd4"] Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.433752 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jn99z/must-gather-dpsd4" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="copy" containerID="cri-o://49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a" gracePeriod=2 Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.458354 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jn99z/must-gather-dpsd4"] Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.924120 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jn99z_must-gather-dpsd4_54433b1a-0157-4690-bffb-01bb356b64fc/copy/0.log" Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.924921 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.937601 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jn99z_must-gather-dpsd4_54433b1a-0157-4690-bffb-01bb356b64fc/copy/0.log" Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.938149 4725 generic.go:334] "Generic (PLEG): container finished" podID="54433b1a-0157-4690-bffb-01bb356b64fc" containerID="49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a" exitCode=143 Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.938209 4725 scope.go:117] "RemoveContainer" containerID="49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a" Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.938276 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jn99z/must-gather-dpsd4" Dec 02 14:06:09 crc kubenswrapper[4725]: I1202 14:06:09.972110 4725 scope.go:117] "RemoveContainer" containerID="90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.031557 4725 scope.go:117] "RemoveContainer" containerID="49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a" Dec 02 14:06:10 crc kubenswrapper[4725]: E1202 14:06:10.032192 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a\": container with ID starting with 49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a not found: ID does not exist" containerID="49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.032245 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a"} err="failed to get container status \"49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a\": rpc error: code = NotFound desc = could not find container \"49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a\": container with ID starting with 49cc2d72001f7dd87757fed8025e6f1f7b46d742575bae4ca774477d47415f2a not found: ID does not exist" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.032274 4725 scope.go:117] "RemoveContainer" containerID="90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7" Dec 02 14:06:10 crc kubenswrapper[4725]: E1202 14:06:10.032639 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7\": container with ID starting with 90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7 not found: ID does not exist" containerID="90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.032665 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7"} err="failed to get container status \"90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7\": rpc error: code = NotFound desc = could not find container \"90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7\": container with ID starting with 90ec72bbb009631cf4b913a5641846ff6a856ecf9899d6177bb7a23809a45ec7 not found: ID does not exist" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.044137 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mq22\" (UniqueName: \"kubernetes.io/projected/54433b1a-0157-4690-bffb-01bb356b64fc-kube-api-access-5mq22\") pod \"54433b1a-0157-4690-bffb-01bb356b64fc\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.044202 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/54433b1a-0157-4690-bffb-01bb356b64fc-must-gather-output\") pod \"54433b1a-0157-4690-bffb-01bb356b64fc\" (UID: \"54433b1a-0157-4690-bffb-01bb356b64fc\") " Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.050171 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54433b1a-0157-4690-bffb-01bb356b64fc-kube-api-access-5mq22" (OuterVolumeSpecName: "kube-api-access-5mq22") pod "54433b1a-0157-4690-bffb-01bb356b64fc" (UID: "54433b1a-0157-4690-bffb-01bb356b64fc"). InnerVolumeSpecName "kube-api-access-5mq22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.147688 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mq22\" (UniqueName: \"kubernetes.io/projected/54433b1a-0157-4690-bffb-01bb356b64fc-kube-api-access-5mq22\") on node \"crc\" DevicePath \"\"" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.194773 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54433b1a-0157-4690-bffb-01bb356b64fc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "54433b1a-0157-4690-bffb-01bb356b64fc" (UID: "54433b1a-0157-4690-bffb-01bb356b64fc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:06:10 crc kubenswrapper[4725]: I1202 14:06:10.249589 4725 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/54433b1a-0157-4690-bffb-01bb356b64fc-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 02 14:06:11 crc kubenswrapper[4725]: I1202 14:06:11.282204 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" path="/var/lib/kubelet/pods/54433b1a-0157-4690-bffb-01bb356b64fc/volumes" Dec 02 14:06:13 crc kubenswrapper[4725]: I1202 14:06:13.268504 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:06:13 crc kubenswrapper[4725]: E1202 14:06:13.269184 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:06:28 crc kubenswrapper[4725]: I1202 14:06:28.268789 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:06:28 crc kubenswrapper[4725]: E1202 14:06:28.269545 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:06:42 crc kubenswrapper[4725]: I1202 14:06:42.269049 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:06:42 crc kubenswrapper[4725]: E1202 14:06:42.269933 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:06:53 crc kubenswrapper[4725]: I1202 14:06:53.268611 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:06:53 crc kubenswrapper[4725]: E1202 14:06:53.269490 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:07:08 crc kubenswrapper[4725]: I1202 14:07:08.268358 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:07:08 crc kubenswrapper[4725]: E1202 14:07:08.269380 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:07:20 crc kubenswrapper[4725]: I1202 14:07:20.268131 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:07:20 crc kubenswrapper[4725]: E1202 14:07:20.268917 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:07:31 crc kubenswrapper[4725]: I1202 14:07:31.268077 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:07:31 crc kubenswrapper[4725]: E1202 14:07:31.268845 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:07:46 crc kubenswrapper[4725]: I1202 14:07:46.268767 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:07:46 crc kubenswrapper[4725]: E1202 14:07:46.270059 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:07:58 crc kubenswrapper[4725]: I1202 14:07:58.275148 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:07:58 crc kubenswrapper[4725]: E1202 14:07:58.276234 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:08:13 crc kubenswrapper[4725]: I1202 14:08:13.268975 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:08:13 crc kubenswrapper[4725]: E1202 14:08:13.269688 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:08:27 crc kubenswrapper[4725]: I1202 14:08:27.269100 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:08:27 crc kubenswrapper[4725]: E1202 14:08:27.269976 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:08:39 crc kubenswrapper[4725]: I1202 14:08:39.275361 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:08:39 crc kubenswrapper[4725]: E1202 14:08:39.276488 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:08:50 crc kubenswrapper[4725]: I1202 14:08:50.268565 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:08:50 crc kubenswrapper[4725]: E1202 14:08:50.269242 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.269124 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.269897 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.478927 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hstpw/must-gather-tl8zg"] Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.479834 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="extract-content" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.479852 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="extract-content" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.479877 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="extract-utilities" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.479884 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="extract-utilities" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.480041 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="extract-content" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.480052 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="extract-content" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.480069 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="copy" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.480075 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="copy" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.480088 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="registry-server" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.480115 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="registry-server" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.480153 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="extract-utilities" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.480159 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="extract-utilities" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.480197 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="registry-server" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.480204 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="registry-server" Dec 02 14:09:01 crc kubenswrapper[4725]: E1202 14:09:01.480231 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="gather" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.480238 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="gather" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.482902 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66f4e11-e42a-4c08-a21a-4d3ea808a89a" containerName="registry-server" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.482923 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="copy" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.482939 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="54433b1a-0157-4690-bffb-01bb356b64fc" containerName="gather" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.482982 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ac3946-a0f4-431e-965b-621442994ad6" containerName="registry-server" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.485441 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.490985 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hstpw"/"default-dockercfg-j97hn" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.492224 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hstpw"/"kube-root-ca.crt" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.498123 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hstpw"/"openshift-service-ca.crt" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.543748 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hstpw/must-gather-tl8zg"] Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.649871 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/eb74a1fa-10e5-4dac-8189-dc56911405cf-must-gather-output\") pod \"must-gather-tl8zg\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.650199 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwzpp\" (UniqueName: \"kubernetes.io/projected/eb74a1fa-10e5-4dac-8189-dc56911405cf-kube-api-access-lwzpp\") pod \"must-gather-tl8zg\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.752375 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/eb74a1fa-10e5-4dac-8189-dc56911405cf-must-gather-output\") pod \"must-gather-tl8zg\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.752858 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/eb74a1fa-10e5-4dac-8189-dc56911405cf-must-gather-output\") pod \"must-gather-tl8zg\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.753530 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwzpp\" (UniqueName: \"kubernetes.io/projected/eb74a1fa-10e5-4dac-8189-dc56911405cf-kube-api-access-lwzpp\") pod \"must-gather-tl8zg\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.773139 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwzpp\" (UniqueName: \"kubernetes.io/projected/eb74a1fa-10e5-4dac-8189-dc56911405cf-kube-api-access-lwzpp\") pod \"must-gather-tl8zg\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:01 crc kubenswrapper[4725]: I1202 14:09:01.820906 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:09:02 crc kubenswrapper[4725]: I1202 14:09:02.358627 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hstpw/must-gather-tl8zg"] Dec 02 14:09:02 crc kubenswrapper[4725]: W1202 14:09:02.366907 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb74a1fa_10e5_4dac_8189_dc56911405cf.slice/crio-c32ee22d20c3fe73cca83a34d9ab7a5a50fda454723b54f9d92723159cb3b049 WatchSource:0}: Error finding container c32ee22d20c3fe73cca83a34d9ab7a5a50fda454723b54f9d92723159cb3b049: Status 404 returned error can't find the container with id c32ee22d20c3fe73cca83a34d9ab7a5a50fda454723b54f9d92723159cb3b049 Dec 02 14:09:02 crc kubenswrapper[4725]: I1202 14:09:02.608889 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/must-gather-tl8zg" event={"ID":"eb74a1fa-10e5-4dac-8189-dc56911405cf","Type":"ContainerStarted","Data":"c32ee22d20c3fe73cca83a34d9ab7a5a50fda454723b54f9d92723159cb3b049"} Dec 02 14:09:03 crc kubenswrapper[4725]: I1202 14:09:03.617716 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/must-gather-tl8zg" event={"ID":"eb74a1fa-10e5-4dac-8189-dc56911405cf","Type":"ContainerStarted","Data":"083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805"} Dec 02 14:09:03 crc kubenswrapper[4725]: I1202 14:09:03.618065 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/must-gather-tl8zg" event={"ID":"eb74a1fa-10e5-4dac-8189-dc56911405cf","Type":"ContainerStarted","Data":"46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7"} Dec 02 14:09:03 crc kubenswrapper[4725]: I1202 14:09:03.634585 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hstpw/must-gather-tl8zg" podStartSLOduration=2.6345643 podStartE2EDuration="2.6345643s" podCreationTimestamp="2025-12-02 14:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 14:09:03.629117115 +0000 UTC m=+3874.585758810" watchObservedRunningTime="2025-12-02 14:09:03.6345643 +0000 UTC m=+3874.591205995" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.296202 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hstpw/crc-debug-r6cb4"] Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.298251 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.451288 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-host\") pod \"crc-debug-r6cb4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.451330 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqcsf\" (UniqueName: \"kubernetes.io/projected/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-kube-api-access-qqcsf\") pod \"crc-debug-r6cb4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.553275 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-host\") pod \"crc-debug-r6cb4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.553323 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqcsf\" (UniqueName: \"kubernetes.io/projected/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-kube-api-access-qqcsf\") pod \"crc-debug-r6cb4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.553443 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-host\") pod \"crc-debug-r6cb4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.573933 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqcsf\" (UniqueName: \"kubernetes.io/projected/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-kube-api-access-qqcsf\") pod \"crc-debug-r6cb4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: I1202 14:09:06.622152 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:06 crc kubenswrapper[4725]: W1202 14:09:06.665894 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1414471_efe5_442a_acc7_a1fa6c4c8ac4.slice/crio-86927e69b083c9ba34dcbf89bba42e812551ebd4f4b3513a7b2b9449aa5a59fc WatchSource:0}: Error finding container 86927e69b083c9ba34dcbf89bba42e812551ebd4f4b3513a7b2b9449aa5a59fc: Status 404 returned error can't find the container with id 86927e69b083c9ba34dcbf89bba42e812551ebd4f4b3513a7b2b9449aa5a59fc Dec 02 14:09:07 crc kubenswrapper[4725]: I1202 14:09:07.651133 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" event={"ID":"c1414471-efe5-442a-acc7-a1fa6c4c8ac4","Type":"ContainerStarted","Data":"f2153dca1e61ea39c8f30acb6f4d884369e8184ff57ed5f2e049c891292fbdb2"} Dec 02 14:09:07 crc kubenswrapper[4725]: I1202 14:09:07.651710 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" event={"ID":"c1414471-efe5-442a-acc7-a1fa6c4c8ac4","Type":"ContainerStarted","Data":"86927e69b083c9ba34dcbf89bba42e812551ebd4f4b3513a7b2b9449aa5a59fc"} Dec 02 14:09:07 crc kubenswrapper[4725]: I1202 14:09:07.670935 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" podStartSLOduration=1.6709163120000001 podStartE2EDuration="1.670916312s" podCreationTimestamp="2025-12-02 14:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 14:09:07.662500773 +0000 UTC m=+3878.619142468" watchObservedRunningTime="2025-12-02 14:09:07.670916312 +0000 UTC m=+3878.627558007" Dec 02 14:09:14 crc kubenswrapper[4725]: I1202 14:09:14.269674 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:09:14 crc kubenswrapper[4725]: E1202 14:09:14.270519 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:09:25 crc kubenswrapper[4725]: I1202 14:09:25.268742 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:09:25 crc kubenswrapper[4725]: E1202 14:09:25.269577 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:09:39 crc kubenswrapper[4725]: I1202 14:09:39.290960 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:09:39 crc kubenswrapper[4725]: E1202 14:09:39.291847 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:09:41 crc kubenswrapper[4725]: I1202 14:09:41.935759 4725 generic.go:334] "Generic (PLEG): container finished" podID="c1414471-efe5-442a-acc7-a1fa6c4c8ac4" containerID="f2153dca1e61ea39c8f30acb6f4d884369e8184ff57ed5f2e049c891292fbdb2" exitCode=0 Dec 02 14:09:41 crc kubenswrapper[4725]: I1202 14:09:41.935844 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" event={"ID":"c1414471-efe5-442a-acc7-a1fa6c4c8ac4","Type":"ContainerDied","Data":"f2153dca1e61ea39c8f30acb6f4d884369e8184ff57ed5f2e049c891292fbdb2"} Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.052479 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.093746 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hstpw/crc-debug-r6cb4"] Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.099351 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqcsf\" (UniqueName: \"kubernetes.io/projected/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-kube-api-access-qqcsf\") pod \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.099448 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-host\") pod \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\" (UID: \"c1414471-efe5-442a-acc7-a1fa6c4c8ac4\") " Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.099898 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-host" (OuterVolumeSpecName: "host") pod "c1414471-efe5-442a-acc7-a1fa6c4c8ac4" (UID: "c1414471-efe5-442a-acc7-a1fa6c4c8ac4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.103976 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hstpw/crc-debug-r6cb4"] Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.109687 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-kube-api-access-qqcsf" (OuterVolumeSpecName: "kube-api-access-qqcsf") pod "c1414471-efe5-442a-acc7-a1fa6c4c8ac4" (UID: "c1414471-efe5-442a-acc7-a1fa6c4c8ac4"). InnerVolumeSpecName "kube-api-access-qqcsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.201566 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqcsf\" (UniqueName: \"kubernetes.io/projected/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-kube-api-access-qqcsf\") on node \"crc\" DevicePath \"\"" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.201598 4725 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c1414471-efe5-442a-acc7-a1fa6c4c8ac4-host\") on node \"crc\" DevicePath \"\"" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.280092 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1414471-efe5-442a-acc7-a1fa6c4c8ac4" path="/var/lib/kubelet/pods/c1414471-efe5-442a-acc7-a1fa6c4c8ac4/volumes" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.954974 4725 scope.go:117] "RemoveContainer" containerID="f2153dca1e61ea39c8f30acb6f4d884369e8184ff57ed5f2e049c891292fbdb2" Dec 02 14:09:43 crc kubenswrapper[4725]: I1202 14:09:43.955020 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-r6cb4" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.275038 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hstpw/crc-debug-rv27w"] Dec 02 14:09:44 crc kubenswrapper[4725]: E1202 14:09:44.275433 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1414471-efe5-442a-acc7-a1fa6c4c8ac4" containerName="container-00" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.275444 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1414471-efe5-442a-acc7-a1fa6c4c8ac4" containerName="container-00" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.275693 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1414471-efe5-442a-acc7-a1fa6c4c8ac4" containerName="container-00" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.276281 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.325030 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4kzb\" (UniqueName: \"kubernetes.io/projected/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-kube-api-access-c4kzb\") pod \"crc-debug-rv27w\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.325105 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-host\") pod \"crc-debug-rv27w\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.426518 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4kzb\" (UniqueName: \"kubernetes.io/projected/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-kube-api-access-c4kzb\") pod \"crc-debug-rv27w\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.426864 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-host\") pod \"crc-debug-rv27w\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.426972 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-host\") pod \"crc-debug-rv27w\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.445826 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4kzb\" (UniqueName: \"kubernetes.io/projected/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-kube-api-access-c4kzb\") pod \"crc-debug-rv27w\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.595283 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.964893 4725 generic.go:334] "Generic (PLEG): container finished" podID="7cadeb1d-a4c9-4f8f-92f4-131e59258da9" containerID="c6d4da9d41206051d88a345c5fdc1f7fbd2746de9da90ee59fb38e3cb47ab38b" exitCode=0 Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.965005 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-rv27w" event={"ID":"7cadeb1d-a4c9-4f8f-92f4-131e59258da9","Type":"ContainerDied","Data":"c6d4da9d41206051d88a345c5fdc1f7fbd2746de9da90ee59fb38e3cb47ab38b"} Dec 02 14:09:44 crc kubenswrapper[4725]: I1202 14:09:44.965310 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-rv27w" event={"ID":"7cadeb1d-a4c9-4f8f-92f4-131e59258da9","Type":"ContainerStarted","Data":"b21047db11f6ac16868e5dd546a323a975c9e6e4b685f360abfd0eb8d92a18d5"} Dec 02 14:09:45 crc kubenswrapper[4725]: I1202 14:09:45.418479 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hstpw/crc-debug-rv27w"] Dec 02 14:09:45 crc kubenswrapper[4725]: I1202 14:09:45.428307 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hstpw/crc-debug-rv27w"] Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.097194 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.156993 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-host\") pod \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.157143 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4kzb\" (UniqueName: \"kubernetes.io/projected/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-kube-api-access-c4kzb\") pod \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\" (UID: \"7cadeb1d-a4c9-4f8f-92f4-131e59258da9\") " Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.157444 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-host" (OuterVolumeSpecName: "host") pod "7cadeb1d-a4c9-4f8f-92f4-131e59258da9" (UID: "7cadeb1d-a4c9-4f8f-92f4-131e59258da9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.157820 4725 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-host\") on node \"crc\" DevicePath \"\"" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.163880 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-kube-api-access-c4kzb" (OuterVolumeSpecName: "kube-api-access-c4kzb") pod "7cadeb1d-a4c9-4f8f-92f4-131e59258da9" (UID: "7cadeb1d-a4c9-4f8f-92f4-131e59258da9"). InnerVolumeSpecName "kube-api-access-c4kzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.262291 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4kzb\" (UniqueName: \"kubernetes.io/projected/7cadeb1d-a4c9-4f8f-92f4-131e59258da9-kube-api-access-c4kzb\") on node \"crc\" DevicePath \"\"" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.586227 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hstpw/crc-debug-q799h"] Dec 02 14:09:46 crc kubenswrapper[4725]: E1202 14:09:46.586658 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cadeb1d-a4c9-4f8f-92f4-131e59258da9" containerName="container-00" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.586671 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cadeb1d-a4c9-4f8f-92f4-131e59258da9" containerName="container-00" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.586902 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cadeb1d-a4c9-4f8f-92f4-131e59258da9" containerName="container-00" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.587560 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.669920 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-host\") pod \"crc-debug-q799h\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.670307 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q4bm\" (UniqueName: \"kubernetes.io/projected/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-kube-api-access-6q4bm\") pod \"crc-debug-q799h\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.772078 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-host\") pod \"crc-debug-q799h\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.772184 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q4bm\" (UniqueName: \"kubernetes.io/projected/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-kube-api-access-6q4bm\") pod \"crc-debug-q799h\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.772249 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-host\") pod \"crc-debug-q799h\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.992277 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b21047db11f6ac16868e5dd546a323a975c9e6e4b685f360abfd0eb8d92a18d5" Dec 02 14:09:46 crc kubenswrapper[4725]: I1202 14:09:46.992341 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-rv27w" Dec 02 14:09:47 crc kubenswrapper[4725]: I1202 14:09:47.282158 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cadeb1d-a4c9-4f8f-92f4-131e59258da9" path="/var/lib/kubelet/pods/7cadeb1d-a4c9-4f8f-92f4-131e59258da9/volumes" Dec 02 14:09:47 crc kubenswrapper[4725]: I1202 14:09:47.290762 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q4bm\" (UniqueName: \"kubernetes.io/projected/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-kube-api-access-6q4bm\") pod \"crc-debug-q799h\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:47 crc kubenswrapper[4725]: I1202 14:09:47.506554 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:47 crc kubenswrapper[4725]: W1202 14:09:47.557750 4725 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90ff4db8_1087_4fb1_bcb8_92f0aba8996b.slice/crio-40bbb97bb21b8c16ff98f9501350715bcb33c25c3edfe2963c2c5f9cb5ca7d1b WatchSource:0}: Error finding container 40bbb97bb21b8c16ff98f9501350715bcb33c25c3edfe2963c2c5f9cb5ca7d1b: Status 404 returned error can't find the container with id 40bbb97bb21b8c16ff98f9501350715bcb33c25c3edfe2963c2c5f9cb5ca7d1b Dec 02 14:09:48 crc kubenswrapper[4725]: I1202 14:09:48.004116 4725 generic.go:334] "Generic (PLEG): container finished" podID="90ff4db8-1087-4fb1-bcb8-92f0aba8996b" containerID="67056639e8095eac580b34d0dd9b59b7fe85fcae24c48d7f474d8fa3fe1b9a0d" exitCode=0 Dec 02 14:09:48 crc kubenswrapper[4725]: I1202 14:09:48.004179 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-q799h" event={"ID":"90ff4db8-1087-4fb1-bcb8-92f0aba8996b","Type":"ContainerDied","Data":"67056639e8095eac580b34d0dd9b59b7fe85fcae24c48d7f474d8fa3fe1b9a0d"} Dec 02 14:09:48 crc kubenswrapper[4725]: I1202 14:09:48.004499 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/crc-debug-q799h" event={"ID":"90ff4db8-1087-4fb1-bcb8-92f0aba8996b","Type":"ContainerStarted","Data":"40bbb97bb21b8c16ff98f9501350715bcb33c25c3edfe2963c2c5f9cb5ca7d1b"} Dec 02 14:09:48 crc kubenswrapper[4725]: I1202 14:09:48.058482 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hstpw/crc-debug-q799h"] Dec 02 14:09:48 crc kubenswrapper[4725]: I1202 14:09:48.070946 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hstpw/crc-debug-q799h"] Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.199336 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sh4j4"] Dec 02 14:09:49 crc kubenswrapper[4725]: E1202 14:09:49.200129 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ff4db8-1087-4fb1-bcb8-92f0aba8996b" containerName="container-00" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.200380 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ff4db8-1087-4fb1-bcb8-92f0aba8996b" containerName="container-00" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.200614 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="90ff4db8-1087-4fb1-bcb8-92f0aba8996b" containerName="container-00" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.202018 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.205413 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sh4j4"] Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.317898 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-catalog-content\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.318019 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbsht\" (UniqueName: \"kubernetes.io/projected/dea78b85-bff2-4868-a817-fb816b6725c8-kube-api-access-bbsht\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.318135 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-utilities\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.420397 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-utilities\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.420522 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-catalog-content\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.420562 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbsht\" (UniqueName: \"kubernetes.io/projected/dea78b85-bff2-4868-a817-fb816b6725c8-kube-api-access-bbsht\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.421153 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-utilities\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.421243 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-catalog-content\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.441768 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbsht\" (UniqueName: \"kubernetes.io/projected/dea78b85-bff2-4868-a817-fb816b6725c8-kube-api-access-bbsht\") pod \"redhat-operators-sh4j4\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.550040 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.641897 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.725450 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q4bm\" (UniqueName: \"kubernetes.io/projected/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-kube-api-access-6q4bm\") pod \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.727065 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-host\") pod \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\" (UID: \"90ff4db8-1087-4fb1-bcb8-92f0aba8996b\") " Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.727603 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-host" (OuterVolumeSpecName: "host") pod "90ff4db8-1087-4fb1-bcb8-92f0aba8996b" (UID: "90ff4db8-1087-4fb1-bcb8-92f0aba8996b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.731961 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-kube-api-access-6q4bm" (OuterVolumeSpecName: "kube-api-access-6q4bm") pod "90ff4db8-1087-4fb1-bcb8-92f0aba8996b" (UID: "90ff4db8-1087-4fb1-bcb8-92f0aba8996b"). InnerVolumeSpecName "kube-api-access-6q4bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.830807 4725 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-host\") on node \"crc\" DevicePath \"\"" Dec 02 14:09:49 crc kubenswrapper[4725]: I1202 14:09:49.831087 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q4bm\" (UniqueName: \"kubernetes.io/projected/90ff4db8-1087-4fb1-bcb8-92f0aba8996b-kube-api-access-6q4bm\") on node \"crc\" DevicePath \"\"" Dec 02 14:09:50 crc kubenswrapper[4725]: I1202 14:09:50.022886 4725 scope.go:117] "RemoveContainer" containerID="67056639e8095eac580b34d0dd9b59b7fe85fcae24c48d7f474d8fa3fe1b9a0d" Dec 02 14:09:50 crc kubenswrapper[4725]: I1202 14:09:50.022920 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/crc-debug-q799h" Dec 02 14:09:50 crc kubenswrapper[4725]: I1202 14:09:50.107197 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sh4j4"] Dec 02 14:09:50 crc kubenswrapper[4725]: I1202 14:09:50.268239 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:09:50 crc kubenswrapper[4725]: E1202 14:09:50.268784 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:09:51 crc kubenswrapper[4725]: I1202 14:09:51.036097 4725 generic.go:334] "Generic (PLEG): container finished" podID="dea78b85-bff2-4868-a817-fb816b6725c8" containerID="f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6" exitCode=0 Dec 02 14:09:51 crc kubenswrapper[4725]: I1202 14:09:51.036292 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerDied","Data":"f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6"} Dec 02 14:09:51 crc kubenswrapper[4725]: I1202 14:09:51.036320 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerStarted","Data":"a3529531113410972ba33910c3974665c06128e29a3f7179cf9de1f1c0870da0"} Dec 02 14:09:51 crc kubenswrapper[4725]: I1202 14:09:51.287658 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90ff4db8-1087-4fb1-bcb8-92f0aba8996b" path="/var/lib/kubelet/pods/90ff4db8-1087-4fb1-bcb8-92f0aba8996b/volumes" Dec 02 14:09:53 crc kubenswrapper[4725]: I1202 14:09:53.055682 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerStarted","Data":"9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc"} Dec 02 14:09:55 crc kubenswrapper[4725]: I1202 14:09:55.076401 4725 generic.go:334] "Generic (PLEG): container finished" podID="dea78b85-bff2-4868-a817-fb816b6725c8" containerID="9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc" exitCode=0 Dec 02 14:09:55 crc kubenswrapper[4725]: I1202 14:09:55.076541 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerDied","Data":"9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc"} Dec 02 14:09:57 crc kubenswrapper[4725]: I1202 14:09:57.096783 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerStarted","Data":"5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a"} Dec 02 14:09:57 crc kubenswrapper[4725]: I1202 14:09:57.115147 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sh4j4" podStartSLOduration=2.925978479 podStartE2EDuration="8.115119891s" podCreationTimestamp="2025-12-02 14:09:49 +0000 UTC" firstStartedPulling="2025-12-02 14:09:51.038643498 +0000 UTC m=+3921.995285193" lastFinishedPulling="2025-12-02 14:09:56.2277849 +0000 UTC m=+3927.184426605" observedRunningTime="2025-12-02 14:09:57.113201023 +0000 UTC m=+3928.069842738" watchObservedRunningTime="2025-12-02 14:09:57.115119891 +0000 UTC m=+3928.071761586" Dec 02 14:09:59 crc kubenswrapper[4725]: I1202 14:09:59.550909 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:09:59 crc kubenswrapper[4725]: I1202 14:09:59.551511 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:10:00 crc kubenswrapper[4725]: I1202 14:10:00.596062 4725 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sh4j4" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="registry-server" probeResult="failure" output=< Dec 02 14:10:00 crc kubenswrapper[4725]: timeout: failed to connect service ":50051" within 1s Dec 02 14:10:00 crc kubenswrapper[4725]: > Dec 02 14:10:04 crc kubenswrapper[4725]: I1202 14:10:04.268762 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:10:05 crc kubenswrapper[4725]: I1202 14:10:05.315136 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"5996f0084ac41bbc5daaf0acbce3974adee0d79875ac90bac80ddb871f615a6d"} Dec 02 14:10:09 crc kubenswrapper[4725]: I1202 14:10:09.598605 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:10:09 crc kubenswrapper[4725]: I1202 14:10:09.670756 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:10:09 crc kubenswrapper[4725]: I1202 14:10:09.832315 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sh4j4"] Dec 02 14:10:11 crc kubenswrapper[4725]: I1202 14:10:11.371071 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sh4j4" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="registry-server" containerID="cri-o://5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a" gracePeriod=2 Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:11.896215 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:11.962241 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-utilities\") pod \"dea78b85-bff2-4868-a817-fb816b6725c8\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:11.962325 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbsht\" (UniqueName: \"kubernetes.io/projected/dea78b85-bff2-4868-a817-fb816b6725c8-kube-api-access-bbsht\") pod \"dea78b85-bff2-4868-a817-fb816b6725c8\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:11.962393 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-catalog-content\") pod \"dea78b85-bff2-4868-a817-fb816b6725c8\" (UID: \"dea78b85-bff2-4868-a817-fb816b6725c8\") " Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:11.963984 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-utilities" (OuterVolumeSpecName: "utilities") pod "dea78b85-bff2-4868-a817-fb816b6725c8" (UID: "dea78b85-bff2-4868-a817-fb816b6725c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:11.970323 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea78b85-bff2-4868-a817-fb816b6725c8-kube-api-access-bbsht" (OuterVolumeSpecName: "kube-api-access-bbsht") pod "dea78b85-bff2-4868-a817-fb816b6725c8" (UID: "dea78b85-bff2-4868-a817-fb816b6725c8"). InnerVolumeSpecName "kube-api-access-bbsht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.064492 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.064791 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbsht\" (UniqueName: \"kubernetes.io/projected/dea78b85-bff2-4868-a817-fb816b6725c8-kube-api-access-bbsht\") on node \"crc\" DevicePath \"\"" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.114701 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dea78b85-bff2-4868-a817-fb816b6725c8" (UID: "dea78b85-bff2-4868-a817-fb816b6725c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.166870 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dea78b85-bff2-4868-a817-fb816b6725c8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.390106 4725 generic.go:334] "Generic (PLEG): container finished" podID="dea78b85-bff2-4868-a817-fb816b6725c8" containerID="5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a" exitCode=0 Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.390157 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerDied","Data":"5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a"} Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.390193 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sh4j4" event={"ID":"dea78b85-bff2-4868-a817-fb816b6725c8","Type":"ContainerDied","Data":"a3529531113410972ba33910c3974665c06128e29a3f7179cf9de1f1c0870da0"} Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.390218 4725 scope.go:117] "RemoveContainer" containerID="5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.390401 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sh4j4" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.421540 4725 scope.go:117] "RemoveContainer" containerID="9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.446728 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sh4j4"] Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.448004 4725 scope.go:117] "RemoveContainer" containerID="f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.470208 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sh4j4"] Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.496676 4725 scope.go:117] "RemoveContainer" containerID="5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a" Dec 02 14:10:12 crc kubenswrapper[4725]: E1202 14:10:12.498964 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a\": container with ID starting with 5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a not found: ID does not exist" containerID="5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.499004 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a"} err="failed to get container status \"5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a\": rpc error: code = NotFound desc = could not find container \"5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a\": container with ID starting with 5a8eba6a2ae5c28c673961311c57e373c47248c0afe57759a703634a1c60993a not found: ID does not exist" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.499025 4725 scope.go:117] "RemoveContainer" containerID="9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc" Dec 02 14:10:12 crc kubenswrapper[4725]: E1202 14:10:12.499329 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc\": container with ID starting with 9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc not found: ID does not exist" containerID="9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.499380 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc"} err="failed to get container status \"9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc\": rpc error: code = NotFound desc = could not find container \"9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc\": container with ID starting with 9cf4116e46dedf73fd55837a4445a0b5ec91f1dd275ac80e23d37b7f05b5a9dc not found: ID does not exist" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.499397 4725 scope.go:117] "RemoveContainer" containerID="f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6" Dec 02 14:10:12 crc kubenswrapper[4725]: E1202 14:10:12.500374 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6\": container with ID starting with f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6 not found: ID does not exist" containerID="f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6" Dec 02 14:10:12 crc kubenswrapper[4725]: I1202 14:10:12.500398 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6"} err="failed to get container status \"f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6\": rpc error: code = NotFound desc = could not find container \"f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6\": container with ID starting with f03035d230798876ce477e6892957dec73f6908c98c7da2f18e6027d9a6dd3c6 not found: ID does not exist" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.166972 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c6d675f88-kk42m_8063dc7f-156e-4f2f-a6fa-8759abb28c37/barbican-api/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.280094 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" path="/var/lib/kubelet/pods/dea78b85-bff2-4868-a817-fb816b6725c8/volumes" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.318020 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c6d675f88-kk42m_8063dc7f-156e-4f2f-a6fa-8759abb28c37/barbican-api-log/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.377075 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54b4d7cbbb-hdq62_e5f2ea09-3acf-4c71-8301-c6531e608bf1/barbican-keystone-listener/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.484579 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-54b4d7cbbb-hdq62_e5f2ea09-3acf-4c71-8301-c6531e608bf1/barbican-keystone-listener-log/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.610854 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c5f8cd4f5-fd8q7_643b3918-6dab-4956-bd36-474ba0195d15/barbican-worker/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.611891 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5c5f8cd4f5-fd8q7_643b3918-6dab-4956-bd36-474ba0195d15/barbican-worker-log/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.772851 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-z8prr_3ab59324-3446-4602-af02-e9094f220f34/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.876413 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/ceilometer-central-agent/0.log" Dec 02 14:10:13 crc kubenswrapper[4725]: I1202 14:10:13.985656 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/proxy-httpd/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.004521 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/ceilometer-notification-agent/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.085972 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c4e7bb4e-6337-40e2-bea3-ef63e192d2c0/sg-core/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.214231 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eda48a5c-28d1-4b25-8377-a8870261551c/cinder-api-log/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.256574 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_eda48a5c-28d1-4b25-8377-a8870261551c/cinder-api/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.462889 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3eab134b-db07-44ea-9f46-734dbda09036/probe/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.520771 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3eab134b-db07-44ea-9f46-734dbda09036/cinder-scheduler/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.637661 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-z7rjc_711edc3c-a714-4e16-a7b8-6f61529f69d7/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.793387 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-d98hr_b013516f-cd7c-4a91-bc3e-bcddd68f439f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:14 crc kubenswrapper[4725]: I1202 14:10:14.871203 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-chnnr_cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61/init/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.083897 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-chnnr_cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61/init/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.134031 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-5m7wg_022c7b77-3848-4ca6-9382-64eb9762bf40/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.225078 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55478c4467-chnnr_cf142ef8-d0b8-440a-ae2c-7f6f0abb7a61/dnsmasq-dns/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.351485 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2b150790-de06-4a61-901f-014c70552a9c/glance-log/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.368474 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2b150790-de06-4a61-901f-014c70552a9c/glance-httpd/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.546121 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17c38050-4a8a-45e9-8ff8-a40dffaee528/glance-httpd/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.577155 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17c38050-4a8a-45e9-8ff8-a40dffaee528/glance-log/0.log" Dec 02 14:10:15 crc kubenswrapper[4725]: I1202 14:10:15.767193 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f9f547644-22mp9_a4cfe8aa-4818-4626-8e86-6607e7bb38d2/horizon/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.197290 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-fc7x8_5fee6c15-b89f-4d6b-8536-0e14c298e235/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.235261 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-bxpzs_375b1a65-e012-4cfa-a08b-5d0b321be9d7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.301893 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f9f547644-22mp9_a4cfe8aa-4818-4626-8e86-6607e7bb38d2/horizon-log/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.534774 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29411401-76w49_7ee9c2e7-85ce-4cfd-9ca0-ebf212087d70/keystone-cron/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.590164 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c7dc85fc9-tj9wl_16fb266f-4c1b-406e-b662-442a92366486/keystone-api/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.775380 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a093384f-2994-4207-8447-1cfe41c8a14e/kube-state-metrics/0.log" Dec 02 14:10:16 crc kubenswrapper[4725]: I1202 14:10:16.853163 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ctnmx_ee65f1b9-bbb7-4561-af64-e6f0ab3deec1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:17 crc kubenswrapper[4725]: I1202 14:10:17.260687 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b5b59474f-qnsqr_e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d/neutron-api/0.log" Dec 02 14:10:17 crc kubenswrapper[4725]: I1202 14:10:17.307369 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b5b59474f-qnsqr_e719e2e0-6c9f-41f0-85c3-6b6e2bfb0b2d/neutron-httpd/0.log" Dec 02 14:10:17 crc kubenswrapper[4725]: I1202 14:10:17.474683 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-g5vgk_d9633180-0a23-47b6-994f-13ec1eed9c30/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.077022 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3aa1cff6-4533-4ba6-aeac-86a813cf8516/nova-api-log/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.189894 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_84138030-0cbd-41bd-8fb4-afe9fcf57bbc/nova-cell0-conductor-conductor/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.466766 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_71998d01-30ab-421a-bddc-5e49c6ac8b29/nova-cell1-conductor-conductor/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.553891 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3aa1cff6-4533-4ba6-aeac-86a813cf8516/nova-api-api/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.631891 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_03aa0ff8-1156-466a-9726-48481ee110d8/nova-cell1-novncproxy-novncproxy/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.759521 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-58vqc_76bb9f49-7e28-49dc-9946-d3e6de9e6a26/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:18 crc kubenswrapper[4725]: I1202 14:10:18.915744 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d/nova-metadata-log/0.log" Dec 02 14:10:19 crc kubenswrapper[4725]: I1202 14:10:19.255847 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc289673-6c5c-45f0-a9ff-df10cddd635c/mysql-bootstrap/0.log" Dec 02 14:10:19 crc kubenswrapper[4725]: I1202 14:10:19.500712 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c1a212a8-1993-4689-988a-b6fc12ad4cb3/nova-scheduler-scheduler/0.log" Dec 02 14:10:19 crc kubenswrapper[4725]: I1202 14:10:19.554441 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc289673-6c5c-45f0-a9ff-df10cddd635c/galera/0.log" Dec 02 14:10:19 crc kubenswrapper[4725]: I1202 14:10:19.566560 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_cc289673-6c5c-45f0-a9ff-df10cddd635c/mysql-bootstrap/0.log" Dec 02 14:10:19 crc kubenswrapper[4725]: I1202 14:10:19.935146 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ef9cb921-2581-495e-be27-cae739b9a971/mysql-bootstrap/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.098040 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ef9cb921-2581-495e-be27-cae739b9a971/mysql-bootstrap/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.183830 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ef9cb921-2581-495e-be27-cae739b9a971/galera/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.366290 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_6ed64991-e56a-4635-b062-f68f7f1d134d/openstackclient/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.407933 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4ngrg_79e8a2ac-c44f-4ca0-b4c4-af57ae34d5b7/ovn-controller/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.457569 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_5b6a3cb0-e50a-4c5b-a83f-87208b3fe04d/nova-metadata-metadata/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.583815 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kwn9x_4739bfa6-9e63-4f3a-bba3-b6c21afe2a74/openstack-network-exporter/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.706404 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovsdb-server-init/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.846982 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovs-vswitchd/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.893323 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovsdb-server-init/0.log" Dec 02 14:10:20 crc kubenswrapper[4725]: I1202 14:10:20.939024 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bvlkl_c552e1c0-2902-4b74-800f-5bccc6ba4427/ovsdb-server/0.log" Dec 02 14:10:21 crc kubenswrapper[4725]: I1202 14:10:21.113485 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-rlrw6_9deb432a-a6c8-4a8d-9b5e-bf50fa45cee1/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:21 crc kubenswrapper[4725]: I1202 14:10:21.172708 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_984ec540-6517-43f8-90e9-751b57ff0165/openstack-network-exporter/0.log" Dec 02 14:10:21 crc kubenswrapper[4725]: I1202 14:10:21.179378 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_984ec540-6517-43f8-90e9-751b57ff0165/ovn-northd/0.log" Dec 02 14:10:21 crc kubenswrapper[4725]: I1202 14:10:21.795477 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bad22ab8-0033-4a28-b7aa-65b1c285a989/openstack-network-exporter/0.log" Dec 02 14:10:21 crc kubenswrapper[4725]: I1202 14:10:21.862444 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bad22ab8-0033-4a28-b7aa-65b1c285a989/ovsdbserver-nb/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.076781 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_51144a39-c6dd-4a5f-adfb-6d62cb459485/openstack-network-exporter/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.147114 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_51144a39-c6dd-4a5f-adfb-6d62cb459485/ovsdbserver-sb/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.278958 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-784f98d458-wsm8f_66a69b36-610f-4f99-b57b-834d0a8a07a1/placement-api/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.435372 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1774e612-2433-4536-a07c-ef5ac77f2036/setup-container/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.435586 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-784f98d458-wsm8f_66a69b36-610f-4f99-b57b-834d0a8a07a1/placement-log/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.662541 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1774e612-2433-4536-a07c-ef5ac77f2036/setup-container/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.731879 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1774e612-2433-4536-a07c-ef5ac77f2036/rabbitmq/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.760734 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7/setup-container/0.log" Dec 02 14:10:22 crc kubenswrapper[4725]: I1202 14:10:22.979693 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7/setup-container/0.log" Dec 02 14:10:23 crc kubenswrapper[4725]: I1202 14:10:23.077857 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-lgzgd_0372a2f9-72d1-468a-bd81-5a1a0a48c545/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:23 crc kubenswrapper[4725]: I1202 14:10:23.084669 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d6f7cbd2-b6c7-4f03-9b6f-dbb7945a5da7/rabbitmq/0.log" Dec 02 14:10:23 crc kubenswrapper[4725]: I1202 14:10:23.839079 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-mjvdg_efdcb239-8753-47be-86b7-267371f72e53/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:23 crc kubenswrapper[4725]: I1202 14:10:23.900944 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vkx8h_8994bc56-f6d8-49d0-8236-eab3fe2d35e7/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.076685 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-twz8w_2c170bed-2666-437d-b8dc-b14bb81bd6eb/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.230788 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9shf4_50bb29f4-5ab8-421c-a5f5-4321d106e1fe/ssh-known-hosts-edpm-deployment/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.362992 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d448f6767-ccmmh_69be1bb2-b1e2-41fb-991c-c6a12db18fb8/proxy-server/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.492838 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6d448f6767-ccmmh_69be1bb2-b1e2-41fb-991c-c6a12db18fb8/proxy-httpd/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.539541 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-z72np_dda4b826-1871-43b5-808d-45d84f2f2483/swift-ring-rebalance/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.699851 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-auditor/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.719945 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-reaper/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.829826 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-replicator/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.893998 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/account-server/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.946671 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-auditor/0.log" Dec 02 14:10:24 crc kubenswrapper[4725]: I1202 14:10:24.975496 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-replicator/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.080353 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-server/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.081639 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/container-updater/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.147775 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-auditor/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.223786 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-expirer/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.297838 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-replicator/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.328120 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-server/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.349599 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/object-updater/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.443643 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/rsync/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.602399 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c3fe5753-c7c5-450a-b4e6-0065deae5cd9/swift-recon-cron/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.625442 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-mhwqc_b45407f7-eb82-40f8-a223-0dfbabc966e5/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.827924 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_564d9d77-b56c-4e22-a4ae-c6fc6f02e7d8/tempest-tests-tempest-tests-runner/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.941253 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_32b05d70-2f30-48dd-8fba-8187969352da/test-operator-logs-container/0.log" Dec 02 14:10:25 crc kubenswrapper[4725]: I1202 14:10:25.996324 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-dgwld_4a420694-62a2-4fd1-aade-ebfaaf73590c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 14:10:32 crc kubenswrapper[4725]: I1202 14:10:32.985074 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_21bbf393-00ed-46c6-a61f-8a458212e8e7/memcached/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.118242 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/util/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.276800 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/pull/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.306361 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/util/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.357810 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/pull/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.523033 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/pull/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.546405 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/util/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.558023 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0c2e265fe131cb14936d72cd088fca614a520bc8ac8ffc13227d716c1067745_856ce13a-68af-4c76-938a-12df01458fa3/extract/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.729413 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ctzjq_b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b/kube-rbac-proxy/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.768274 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-z9ttg_570a3a73-dffb-4ded-8fd1-d30ddc60979d/kube-rbac-proxy/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.787813 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-ctzjq_b128b4b7-4c22-4ed3-ac83-ed8ca3c0508b/manager/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.923022 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-z9ttg_570a3a73-dffb-4ded-8fd1-d30ddc60979d/manager/0.log" Dec 02 14:10:53 crc kubenswrapper[4725]: I1202 14:10:53.978700 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-ppgnw_c42118d3-322d-4733-8749-d13bc174a6f8/kube-rbac-proxy/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.037571 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-ppgnw_c42118d3-322d-4733-8749-d13bc174a6f8/manager/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.165148 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-njlj7_65a9f532-387a-487e-bfe0-81a110509366/kube-rbac-proxy/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.286897 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-njlj7_65a9f532-387a-487e-bfe0-81a110509366/manager/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.382562 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cwfk4_0b562f2a-162c-4611-81d3-00937e945f96/kube-rbac-proxy/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.389745 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-cwfk4_0b562f2a-162c-4611-81d3-00937e945f96/manager/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.549048 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pmdw6_6b08a8fe-08d5-4b04-9283-51973a9ac3d6/kube-rbac-proxy/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.598785 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-pmdw6_6b08a8fe-08d5-4b04-9283-51973a9ac3d6/manager/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.722348 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g99db_bf404326-3c43-4cc4-93b7-0793213afcba/kube-rbac-proxy/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.907346 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-klrxl_4bad83be-e3ed-499a-8e64-cfe83c4ac1e6/kube-rbac-proxy/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.914795 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g99db_bf404326-3c43-4cc4-93b7-0793213afcba/manager/0.log" Dec 02 14:10:54 crc kubenswrapper[4725]: I1202 14:10:54.984085 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-klrxl_4bad83be-e3ed-499a-8e64-cfe83c4ac1e6/manager/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.120206 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p22fj_3fb7e977-cd14-4b71-b349-19d4487cfb15/kube-rbac-proxy/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.146440 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p22fj_3fb7e977-cd14-4b71-b349-19d4487cfb15/manager/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.298916 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rk99q_c1f336f6-716b-4096-89cd-f23d1fed90cb/kube-rbac-proxy/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.329524 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-rk99q_c1f336f6-716b-4096-89cd-f23d1fed90cb/manager/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.438896 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-8s44s_8f32e184-607a-4077-aa41-ca7aabba76aa/kube-rbac-proxy/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.525543 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-8s44s_8f32e184-607a-4077-aa41-ca7aabba76aa/manager/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.561231 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-wd49v_2c988ae2-9613-4584-a599-a7d063611189/kube-rbac-proxy/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.686749 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-wd49v_2c988ae2-9613-4584-a599-a7d063611189/manager/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.746312 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-cllpp_eacfb091-583b-4e89-9850-cc146cef7e20/kube-rbac-proxy/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.829380 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-cllpp_eacfb091-583b-4e89-9850-cc146cef7e20/manager/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.925310 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-zdj64_bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b/kube-rbac-proxy/0.log" Dec 02 14:10:55 crc kubenswrapper[4725]: I1202 14:10:55.982130 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-zdj64_bd2ee9ec-8397-4e64-8ea9-dc8f0466ab4b/manager/0.log" Dec 02 14:10:56 crc kubenswrapper[4725]: I1202 14:10:56.109187 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc_9c4065a8-50f2-4be9-bdd7-42345814f4c5/kube-rbac-proxy/0.log" Dec 02 14:10:56 crc kubenswrapper[4725]: I1202 14:10:56.149017 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q4ddc_9c4065a8-50f2-4be9-bdd7-42345814f4c5/manager/0.log" Dec 02 14:10:56 crc kubenswrapper[4725]: I1202 14:10:56.561338 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-75ff9f7b7-wcxkf_98fc37a5-6d29-4f0b-93ab-ffecc2157b33/operator/0.log" Dec 02 14:10:56 crc kubenswrapper[4725]: I1202 14:10:56.793107 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-59k9r_6f802fa6-11b4-40ef-b9f1-bfb654645474/registry-server/0.log" Dec 02 14:10:56 crc kubenswrapper[4725]: I1202 14:10:56.984437 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mxpqd_511bab3c-3885-4c5a-8c91-bd231b1c2a03/kube-rbac-proxy/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.141792 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mxpqd_511bab3c-3885-4c5a-8c91-bd231b1c2a03/manager/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.190967 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wqn2j_ad9c89d4-9792-401b-9f1f-81d2e5c7ede9/kube-rbac-proxy/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.258582 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wqn2j_ad9c89d4-9792-401b-9f1f-81d2e5c7ede9/manager/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.398395 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-754459f956-vkwfn_e68bdc86-d09e-48ae-839b-9866e70e1383/manager/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.446765 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-ctlsq_6b03a06a-b4d5-4443-98eb-fdea3fd8302c/operator/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.542813 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-cptsj_dc849752-f58a-4e7a-b1fa-5b7d88c14d1a/kube-rbac-proxy/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.598759 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-cptsj_dc849752-f58a-4e7a-b1fa-5b7d88c14d1a/manager/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.703719 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-2l2x4_bd74c0bd-9ef5-487c-b190-2b23a8120c8d/kube-rbac-proxy/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.704129 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-2l2x4_bd74c0bd-9ef5-487c-b190-2b23a8120c8d/manager/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.793686 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-nrrpt_7965a5b4-89be-44c6-8471-20ede0f6c92f/kube-rbac-proxy/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.855423 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-nrrpt_7965a5b4-89be-44c6-8471-20ede0f6c92f/manager/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.928860 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-rrnmn_b71451df-d329-44b4-8d63-f0ea35db6ce7/kube-rbac-proxy/0.log" Dec 02 14:10:57 crc kubenswrapper[4725]: I1202 14:10:57.994488 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-rrnmn_b71451df-d329-44b4-8d63-f0ea35db6ce7/manager/0.log" Dec 02 14:11:16 crc kubenswrapper[4725]: I1202 14:11:16.444784 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bphpt_4d37326e-fdd1-41e0-8401-5f47038356b3/control-plane-machine-set-operator/0.log" Dec 02 14:11:16 crc kubenswrapper[4725]: I1202 14:11:16.625597 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h8fhj_c17fbfab-7855-44b3-a983-22915062793a/kube-rbac-proxy/0.log" Dec 02 14:11:16 crc kubenswrapper[4725]: I1202 14:11:16.689316 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-h8fhj_c17fbfab-7855-44b3-a983-22915062793a/machine-api-operator/0.log" Dec 02 14:11:28 crc kubenswrapper[4725]: I1202 14:11:28.304292 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6prlb_81141996-2325-4372-81be-7f69cbec7194/cert-manager-controller/0.log" Dec 02 14:11:28 crc kubenswrapper[4725]: I1202 14:11:28.489193 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-f864q_864d4bd0-504a-4d91-bb28-b920fe8845aa/cert-manager-cainjector/0.log" Dec 02 14:11:28 crc kubenswrapper[4725]: I1202 14:11:28.581572 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-hsrpm_f84fbbc8-bcb0-4aff-b5a4-48e932ec4ac4/cert-manager-webhook/0.log" Dec 02 14:11:39 crc kubenswrapper[4725]: I1202 14:11:39.976728 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5dmqd_a4844270-686d-4f0b-aafc-5b2b1924444f/nmstate-console-plugin/0.log" Dec 02 14:11:40 crc kubenswrapper[4725]: I1202 14:11:40.123991 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6cqs8_28a8d066-b52d-450e-a738-884b8b260da5/nmstate-handler/0.log" Dec 02 14:11:40 crc kubenswrapper[4725]: I1202 14:11:40.144302 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-v7ngp_93447f0a-ea95-4ec5-beea-8733e261d5f3/kube-rbac-proxy/0.log" Dec 02 14:11:40 crc kubenswrapper[4725]: I1202 14:11:40.179063 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-v7ngp_93447f0a-ea95-4ec5-beea-8733e261d5f3/nmstate-metrics/0.log" Dec 02 14:11:40 crc kubenswrapper[4725]: I1202 14:11:40.326289 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-xz2lq_511d65e3-135b-4218-aaeb-bee781e9befb/nmstate-operator/0.log" Dec 02 14:11:40 crc kubenswrapper[4725]: I1202 14:11:40.398643 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-9rkbp_15ebd594-c399-41a5-a80a-6ce365f9e657/nmstate-webhook/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.233513 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bkrzn_ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b/kube-rbac-proxy/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.233770 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-bkrzn_ff86b6f3-7d02-4d1e-b212-7d97b47f0e5b/controller/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.399906 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.627905 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.635396 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.647535 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.647608 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.816294 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.857076 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.859932 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.874812 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:11:55 crc kubenswrapper[4725]: I1202 14:11:55.998143 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-frr-files/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.023526 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-metrics/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.055402 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/cp-reloader/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.058046 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/controller/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.775101 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/kube-rbac-proxy/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.780377 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/frr-metrics/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.789021 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/kube-rbac-proxy-frr/0.log" Dec 02 14:11:56 crc kubenswrapper[4725]: I1202 14:11:56.987855 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/reloader/0.log" Dec 02 14:11:57 crc kubenswrapper[4725]: I1202 14:11:57.050873 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-c7g46_00168d15-ce5e-4164-891c-1403031bee1d/frr-k8s-webhook-server/0.log" Dec 02 14:11:57 crc kubenswrapper[4725]: I1202 14:11:57.337890 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-747844468f-qq5ns_bc5b46e6-ecea-4260-a386-e7408c29a842/manager/0.log" Dec 02 14:11:57 crc kubenswrapper[4725]: I1202 14:11:57.446081 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-857fd986b8-tqd7q_2315c4fc-8de9-4d85-9fab-a2cedd0eb95d/webhook-server/0.log" Dec 02 14:11:57 crc kubenswrapper[4725]: I1202 14:11:57.561316 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dhmrd_ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb/kube-rbac-proxy/0.log" Dec 02 14:11:58 crc kubenswrapper[4725]: I1202 14:11:58.094868 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dhmrd_ee868a5e-5d4c-49c8-a7dc-43368d3cf5fb/speaker/0.log" Dec 02 14:11:58 crc kubenswrapper[4725]: I1202 14:11:58.323549 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-bspx6_46316422-5e47-461c-8827-6e12d896134c/frr/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.494114 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/util/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.693278 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/util/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.715034 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/pull/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.748783 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/pull/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.869804 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/pull/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.897956 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/util/0.log" Dec 02 14:12:09 crc kubenswrapper[4725]: I1202 14:12:09.897969 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f9x8hk_c0ff45e5-0175-458b-a1e2-a4079690d9b0/extract/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.069452 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/util/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.240429 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/pull/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.253849 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/pull/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.267114 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/util/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.414770 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/util/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.431101 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/pull/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.498622 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83h8ktr_a9565732-da1b-4ccc-8913-fd706474ce9a/extract/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.602735 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-utilities/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.767137 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-utilities/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.789786 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-content/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.793446 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-content/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.946893 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-utilities/0.log" Dec 02 14:12:10 crc kubenswrapper[4725]: I1202 14:12:10.973693 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/extract-content/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.160249 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-utilities/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.412314 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-content/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.426285 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-utilities/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.463754 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-content/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.553399 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rs8r5_38d5c377-0464-41bd-aff4-de31874ea6b9/registry-server/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.653270 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-content/0.log" Dec 02 14:12:11 crc kubenswrapper[4725]: I1202 14:12:11.716485 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/extract-utilities/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.007833 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kcshx_55b1c898-ab77-4a63-ba54-4c3246ec5732/marketplace-operator/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.112525 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-utilities/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.332856 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-utilities/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.337285 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-content/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.393231 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-content/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.438661 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-nbhnj_f7a00dd9-d19f-415f-bad6-0114f989bdae/registry-server/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.569623 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-content/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.571038 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/extract-utilities/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.760635 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ss4h6_1c2ff71e-0534-496d-8c56-2413e0352da4/registry-server/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.765514 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-utilities/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.919962 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-content/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.937614 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-utilities/0.log" Dec 02 14:12:12 crc kubenswrapper[4725]: I1202 14:12:12.955398 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-content/0.log" Dec 02 14:12:13 crc kubenswrapper[4725]: I1202 14:12:13.142844 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-utilities/0.log" Dec 02 14:12:13 crc kubenswrapper[4725]: I1202 14:12:13.180971 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/extract-content/0.log" Dec 02 14:12:13 crc kubenswrapper[4725]: I1202 14:12:13.564303 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4qzhx_ea32e5f6-b2b8-49a4-b7a0-2670dd62561b/registry-server/0.log" Dec 02 14:12:24 crc kubenswrapper[4725]: I1202 14:12:24.284401 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:12:24 crc kubenswrapper[4725]: I1202 14:12:24.285203 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:12:54 crc kubenswrapper[4725]: I1202 14:12:54.285091 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:12:54 crc kubenswrapper[4725]: I1202 14:12:54.285710 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:13:24 crc kubenswrapper[4725]: I1202 14:13:24.284195 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:13:24 crc kubenswrapper[4725]: I1202 14:13:24.284714 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:13:24 crc kubenswrapper[4725]: I1202 14:13:24.284757 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 14:13:24 crc kubenswrapper[4725]: I1202 14:13:24.285415 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5996f0084ac41bbc5daaf0acbce3974adee0d79875ac90bac80ddb871f615a6d"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 14:13:24 crc kubenswrapper[4725]: I1202 14:13:24.285478 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://5996f0084ac41bbc5daaf0acbce3974adee0d79875ac90bac80ddb871f615a6d" gracePeriod=600 Dec 02 14:13:25 crc kubenswrapper[4725]: I1202 14:13:25.203266 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="5996f0084ac41bbc5daaf0acbce3974adee0d79875ac90bac80ddb871f615a6d" exitCode=0 Dec 02 14:13:25 crc kubenswrapper[4725]: I1202 14:13:25.203493 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"5996f0084ac41bbc5daaf0acbce3974adee0d79875ac90bac80ddb871f615a6d"} Dec 02 14:13:25 crc kubenswrapper[4725]: I1202 14:13:25.203810 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerStarted","Data":"669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438"} Dec 02 14:13:25 crc kubenswrapper[4725]: I1202 14:13:25.203845 4725 scope.go:117] "RemoveContainer" containerID="0471e38ad2a1983bbc025f4778c750ad78c489deb60391c8e004644a56829376" Dec 02 14:13:54 crc kubenswrapper[4725]: I1202 14:13:54.456245 4725 generic.go:334] "Generic (PLEG): container finished" podID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerID="46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7" exitCode=0 Dec 02 14:13:54 crc kubenswrapper[4725]: I1202 14:13:54.456354 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hstpw/must-gather-tl8zg" event={"ID":"eb74a1fa-10e5-4dac-8189-dc56911405cf","Type":"ContainerDied","Data":"46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7"} Dec 02 14:13:54 crc kubenswrapper[4725]: I1202 14:13:54.459087 4725 scope.go:117] "RemoveContainer" containerID="46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7" Dec 02 14:13:55 crc kubenswrapper[4725]: I1202 14:13:55.484827 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hstpw_must-gather-tl8zg_eb74a1fa-10e5-4dac-8189-dc56911405cf/gather/0.log" Dec 02 14:14:05 crc kubenswrapper[4725]: I1202 14:14:05.899218 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hstpw/must-gather-tl8zg"] Dec 02 14:14:05 crc kubenswrapper[4725]: I1202 14:14:05.899909 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hstpw/must-gather-tl8zg" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="copy" containerID="cri-o://083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805" gracePeriod=2 Dec 02 14:14:05 crc kubenswrapper[4725]: I1202 14:14:05.908550 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hstpw/must-gather-tl8zg"] Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.306141 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hstpw_must-gather-tl8zg_eb74a1fa-10e5-4dac-8189-dc56911405cf/copy/0.log" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.307029 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.364617 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwzpp\" (UniqueName: \"kubernetes.io/projected/eb74a1fa-10e5-4dac-8189-dc56911405cf-kube-api-access-lwzpp\") pod \"eb74a1fa-10e5-4dac-8189-dc56911405cf\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.364809 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/eb74a1fa-10e5-4dac-8189-dc56911405cf-must-gather-output\") pod \"eb74a1fa-10e5-4dac-8189-dc56911405cf\" (UID: \"eb74a1fa-10e5-4dac-8189-dc56911405cf\") " Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.380709 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb74a1fa-10e5-4dac-8189-dc56911405cf-kube-api-access-lwzpp" (OuterVolumeSpecName: "kube-api-access-lwzpp") pod "eb74a1fa-10e5-4dac-8189-dc56911405cf" (UID: "eb74a1fa-10e5-4dac-8189-dc56911405cf"). InnerVolumeSpecName "kube-api-access-lwzpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.467205 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwzpp\" (UniqueName: \"kubernetes.io/projected/eb74a1fa-10e5-4dac-8189-dc56911405cf-kube-api-access-lwzpp\") on node \"crc\" DevicePath \"\"" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.531162 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb74a1fa-10e5-4dac-8189-dc56911405cf-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "eb74a1fa-10e5-4dac-8189-dc56911405cf" (UID: "eb74a1fa-10e5-4dac-8189-dc56911405cf"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.569161 4725 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/eb74a1fa-10e5-4dac-8189-dc56911405cf-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.570616 4725 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hstpw_must-gather-tl8zg_eb74a1fa-10e5-4dac-8189-dc56911405cf/copy/0.log" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.570988 4725 generic.go:334] "Generic (PLEG): container finished" podID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerID="083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805" exitCode=143 Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.571041 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hstpw/must-gather-tl8zg" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.571053 4725 scope.go:117] "RemoveContainer" containerID="083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.600884 4725 scope.go:117] "RemoveContainer" containerID="46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.681419 4725 scope.go:117] "RemoveContainer" containerID="083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805" Dec 02 14:14:06 crc kubenswrapper[4725]: E1202 14:14:06.682081 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805\": container with ID starting with 083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805 not found: ID does not exist" containerID="083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.682113 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805"} err="failed to get container status \"083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805\": rpc error: code = NotFound desc = could not find container \"083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805\": container with ID starting with 083e2bb609a4b76e5fb82f266e4ca7e8d73337c8305506bba5f664c9339ad805 not found: ID does not exist" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.682134 4725 scope.go:117] "RemoveContainer" containerID="46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7" Dec 02 14:14:06 crc kubenswrapper[4725]: E1202 14:14:06.682614 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7\": container with ID starting with 46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7 not found: ID does not exist" containerID="46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7" Dec 02 14:14:06 crc kubenswrapper[4725]: I1202 14:14:06.682655 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7"} err="failed to get container status \"46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7\": rpc error: code = NotFound desc = could not find container \"46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7\": container with ID starting with 46edcefa58f643eac532c804a58ec19dc1ecd0e2a04a7fd44fa2375d075bcfe7 not found: ID does not exist" Dec 02 14:14:07 crc kubenswrapper[4725]: I1202 14:14:07.278525 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" path="/var/lib/kubelet/pods/eb74a1fa-10e5-4dac-8189-dc56911405cf/volumes" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.180201 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l"] Dec 02 14:15:00 crc kubenswrapper[4725]: E1202 14:15:00.181272 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="extract-utilities" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181287 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="extract-utilities" Dec 02 14:15:00 crc kubenswrapper[4725]: E1202 14:15:00.181305 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="registry-server" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181311 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="registry-server" Dec 02 14:15:00 crc kubenswrapper[4725]: E1202 14:15:00.181322 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="gather" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181327 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="gather" Dec 02 14:15:00 crc kubenswrapper[4725]: E1202 14:15:00.181352 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="extract-content" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181358 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="extract-content" Dec 02 14:15:00 crc kubenswrapper[4725]: E1202 14:15:00.181376 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="copy" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181381 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="copy" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181656 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea78b85-bff2-4868-a817-fb816b6725c8" containerName="registry-server" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181678 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="gather" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.181692 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb74a1fa-10e5-4dac-8189-dc56911405cf" containerName="copy" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.182542 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.185484 4725 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.185710 4725 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.192108 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l"] Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.332092 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea46708-80fe-4870-bd7c-c920648859d8-secret-volume\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.332186 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea46708-80fe-4870-bd7c-c920648859d8-config-volume\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.332248 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxh8f\" (UniqueName: \"kubernetes.io/projected/2ea46708-80fe-4870-bd7c-c920648859d8-kube-api-access-dxh8f\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.433913 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea46708-80fe-4870-bd7c-c920648859d8-config-volume\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.434012 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxh8f\" (UniqueName: \"kubernetes.io/projected/2ea46708-80fe-4870-bd7c-c920648859d8-kube-api-access-dxh8f\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.434131 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea46708-80fe-4870-bd7c-c920648859d8-secret-volume\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.435292 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea46708-80fe-4870-bd7c-c920648859d8-config-volume\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.440769 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea46708-80fe-4870-bd7c-c920648859d8-secret-volume\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.457662 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxh8f\" (UniqueName: \"kubernetes.io/projected/2ea46708-80fe-4870-bd7c-c920648859d8-kube-api-access-dxh8f\") pod \"collect-profiles-29411415-zcs2l\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.528913 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:00 crc kubenswrapper[4725]: I1202 14:15:00.988131 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l"] Dec 02 14:15:01 crc kubenswrapper[4725]: I1202 14:15:01.047398 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" event={"ID":"2ea46708-80fe-4870-bd7c-c920648859d8","Type":"ContainerStarted","Data":"e63a7a4b6f225e153af7e53a44dfc2720d9dc43798159f9aaf1c06b7103c23b9"} Dec 02 14:15:02 crc kubenswrapper[4725]: I1202 14:15:02.057294 4725 generic.go:334] "Generic (PLEG): container finished" podID="2ea46708-80fe-4870-bd7c-c920648859d8" containerID="71ed6288daf9cd82d1c3bbb88755295ea885144b4ed92ed40b53dd6166e93449" exitCode=0 Dec 02 14:15:02 crc kubenswrapper[4725]: I1202 14:15:02.057527 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" event={"ID":"2ea46708-80fe-4870-bd7c-c920648859d8","Type":"ContainerDied","Data":"71ed6288daf9cd82d1c3bbb88755295ea885144b4ed92ed40b53dd6166e93449"} Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.397333 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.594637 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea46708-80fe-4870-bd7c-c920648859d8-config-volume\") pod \"2ea46708-80fe-4870-bd7c-c920648859d8\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.594731 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxh8f\" (UniqueName: \"kubernetes.io/projected/2ea46708-80fe-4870-bd7c-c920648859d8-kube-api-access-dxh8f\") pod \"2ea46708-80fe-4870-bd7c-c920648859d8\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.594908 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea46708-80fe-4870-bd7c-c920648859d8-secret-volume\") pod \"2ea46708-80fe-4870-bd7c-c920648859d8\" (UID: \"2ea46708-80fe-4870-bd7c-c920648859d8\") " Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.595490 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea46708-80fe-4870-bd7c-c920648859d8-config-volume" (OuterVolumeSpecName: "config-volume") pod "2ea46708-80fe-4870-bd7c-c920648859d8" (UID: "2ea46708-80fe-4870-bd7c-c920648859d8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.600725 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea46708-80fe-4870-bd7c-c920648859d8-kube-api-access-dxh8f" (OuterVolumeSpecName: "kube-api-access-dxh8f") pod "2ea46708-80fe-4870-bd7c-c920648859d8" (UID: "2ea46708-80fe-4870-bd7c-c920648859d8"). InnerVolumeSpecName "kube-api-access-dxh8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.605626 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea46708-80fe-4870-bd7c-c920648859d8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2ea46708-80fe-4870-bd7c-c920648859d8" (UID: "2ea46708-80fe-4870-bd7c-c920648859d8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.696888 4725 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ea46708-80fe-4870-bd7c-c920648859d8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.696935 4725 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ea46708-80fe-4870-bd7c-c920648859d8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 14:15:03 crc kubenswrapper[4725]: I1202 14:15:03.696951 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxh8f\" (UniqueName: \"kubernetes.io/projected/2ea46708-80fe-4870-bd7c-c920648859d8-kube-api-access-dxh8f\") on node \"crc\" DevicePath \"\"" Dec 02 14:15:04 crc kubenswrapper[4725]: I1202 14:15:04.076409 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" event={"ID":"2ea46708-80fe-4870-bd7c-c920648859d8","Type":"ContainerDied","Data":"e63a7a4b6f225e153af7e53a44dfc2720d9dc43798159f9aaf1c06b7103c23b9"} Dec 02 14:15:04 crc kubenswrapper[4725]: I1202 14:15:04.076690 4725 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e63a7a4b6f225e153af7e53a44dfc2720d9dc43798159f9aaf1c06b7103c23b9" Dec 02 14:15:04 crc kubenswrapper[4725]: I1202 14:15:04.076493 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29411415-zcs2l" Dec 02 14:15:04 crc kubenswrapper[4725]: I1202 14:15:04.476280 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz"] Dec 02 14:15:04 crc kubenswrapper[4725]: I1202 14:15:04.484497 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29411370-2x9tz"] Dec 02 14:15:05 crc kubenswrapper[4725]: I1202 14:15:05.279386 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c51f8b2-f00f-47e4-92ae-000fac82e331" path="/var/lib/kubelet/pods/5c51f8b2-f00f-47e4-92ae-000fac82e331/volumes" Dec 02 14:15:24 crc kubenswrapper[4725]: I1202 14:15:24.284353 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:15:24 crc kubenswrapper[4725]: I1202 14:15:24.285026 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:15:51 crc kubenswrapper[4725]: I1202 14:15:51.253273 4725 scope.go:117] "RemoveContainer" containerID="c6d4da9d41206051d88a345c5fdc1f7fbd2746de9da90ee59fb38e3cb47ab38b" Dec 02 14:15:51 crc kubenswrapper[4725]: I1202 14:15:51.283890 4725 scope.go:117] "RemoveContainer" containerID="93ba1471b4e1675c3bf7158abf29ff66ee0f802671eb26a8b191dcfa81f7c4d1" Dec 02 14:15:54 crc kubenswrapper[4725]: I1202 14:15:54.284628 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:15:54 crc kubenswrapper[4725]: I1202 14:15:54.285882 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:15:58 crc kubenswrapper[4725]: I1202 14:15:58.989728 4725 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pllrb"] Dec 02 14:15:58 crc kubenswrapper[4725]: E1202 14:15:58.990713 4725 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea46708-80fe-4870-bd7c-c920648859d8" containerName="collect-profiles" Dec 02 14:15:58 crc kubenswrapper[4725]: I1202 14:15:58.990728 4725 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea46708-80fe-4870-bd7c-c920648859d8" containerName="collect-profiles" Dec 02 14:15:58 crc kubenswrapper[4725]: I1202 14:15:58.990967 4725 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea46708-80fe-4870-bd7c-c920648859d8" containerName="collect-profiles" Dec 02 14:15:58 crc kubenswrapper[4725]: I1202 14:15:58.992595 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.008034 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pllrb"] Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.087806 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qshn\" (UniqueName: \"kubernetes.io/projected/4e4c2870-ab09-4794-b3c9-836df98015ad-kube-api-access-8qshn\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.087863 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-catalog-content\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.088152 4725 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-utilities\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.189828 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qshn\" (UniqueName: \"kubernetes.io/projected/4e4c2870-ab09-4794-b3c9-836df98015ad-kube-api-access-8qshn\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.189902 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-catalog-content\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.190035 4725 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-utilities\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.190808 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-catalog-content\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.190895 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-utilities\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.213886 4725 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qshn\" (UniqueName: \"kubernetes.io/projected/4e4c2870-ab09-4794-b3c9-836df98015ad-kube-api-access-8qshn\") pod \"redhat-marketplace-pllrb\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.324044 4725 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:15:59 crc kubenswrapper[4725]: I1202 14:15:59.878632 4725 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pllrb"] Dec 02 14:16:00 crc kubenswrapper[4725]: I1202 14:16:00.592370 4725 generic.go:334] "Generic (PLEG): container finished" podID="4e4c2870-ab09-4794-b3c9-836df98015ad" containerID="1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e" exitCode=0 Dec 02 14:16:00 crc kubenswrapper[4725]: I1202 14:16:00.592435 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pllrb" event={"ID":"4e4c2870-ab09-4794-b3c9-836df98015ad","Type":"ContainerDied","Data":"1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e"} Dec 02 14:16:00 crc kubenswrapper[4725]: I1202 14:16:00.592719 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pllrb" event={"ID":"4e4c2870-ab09-4794-b3c9-836df98015ad","Type":"ContainerStarted","Data":"3c2abdf367e1fbe34a33e1dd4089ec679ff9e7f336e8b789a970e8d2ff7ff6b6"} Dec 02 14:16:00 crc kubenswrapper[4725]: I1202 14:16:00.594941 4725 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 14:16:02 crc kubenswrapper[4725]: I1202 14:16:02.613924 4725 generic.go:334] "Generic (PLEG): container finished" podID="4e4c2870-ab09-4794-b3c9-836df98015ad" containerID="3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d" exitCode=0 Dec 02 14:16:02 crc kubenswrapper[4725]: I1202 14:16:02.614019 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pllrb" event={"ID":"4e4c2870-ab09-4794-b3c9-836df98015ad","Type":"ContainerDied","Data":"3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d"} Dec 02 14:16:03 crc kubenswrapper[4725]: I1202 14:16:03.624584 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pllrb" event={"ID":"4e4c2870-ab09-4794-b3c9-836df98015ad","Type":"ContainerStarted","Data":"a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602"} Dec 02 14:16:03 crc kubenswrapper[4725]: I1202 14:16:03.641892 4725 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pllrb" podStartSLOduration=2.93664303 podStartE2EDuration="5.641877769s" podCreationTimestamp="2025-12-02 14:15:58 +0000 UTC" firstStartedPulling="2025-12-02 14:16:00.594723125 +0000 UTC m=+4291.551364820" lastFinishedPulling="2025-12-02 14:16:03.299957854 +0000 UTC m=+4294.256599559" observedRunningTime="2025-12-02 14:16:03.638651808 +0000 UTC m=+4294.595293503" watchObservedRunningTime="2025-12-02 14:16:03.641877769 +0000 UTC m=+4294.598519464" Dec 02 14:16:09 crc kubenswrapper[4725]: I1202 14:16:09.324384 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:16:09 crc kubenswrapper[4725]: I1202 14:16:09.324984 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:16:09 crc kubenswrapper[4725]: I1202 14:16:09.368145 4725 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:16:09 crc kubenswrapper[4725]: I1202 14:16:09.715973 4725 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:16:09 crc kubenswrapper[4725]: I1202 14:16:09.759427 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pllrb"] Dec 02 14:16:11 crc kubenswrapper[4725]: I1202 14:16:11.692068 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pllrb" podUID="4e4c2870-ab09-4794-b3c9-836df98015ad" containerName="registry-server" containerID="cri-o://a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602" gracePeriod=2 Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.147945 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.344152 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-utilities\") pod \"4e4c2870-ab09-4794-b3c9-836df98015ad\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.344322 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qshn\" (UniqueName: \"kubernetes.io/projected/4e4c2870-ab09-4794-b3c9-836df98015ad-kube-api-access-8qshn\") pod \"4e4c2870-ab09-4794-b3c9-836df98015ad\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.344402 4725 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-catalog-content\") pod \"4e4c2870-ab09-4794-b3c9-836df98015ad\" (UID: \"4e4c2870-ab09-4794-b3c9-836df98015ad\") " Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.346019 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-utilities" (OuterVolumeSpecName: "utilities") pod "4e4c2870-ab09-4794-b3c9-836df98015ad" (UID: "4e4c2870-ab09-4794-b3c9-836df98015ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.350422 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e4c2870-ab09-4794-b3c9-836df98015ad-kube-api-access-8qshn" (OuterVolumeSpecName: "kube-api-access-8qshn") pod "4e4c2870-ab09-4794-b3c9-836df98015ad" (UID: "4e4c2870-ab09-4794-b3c9-836df98015ad"). InnerVolumeSpecName "kube-api-access-8qshn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.368284 4725 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e4c2870-ab09-4794-b3c9-836df98015ad" (UID: "4e4c2870-ab09-4794-b3c9-836df98015ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.446145 4725 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.446355 4725 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qshn\" (UniqueName: \"kubernetes.io/projected/4e4c2870-ab09-4794-b3c9-836df98015ad-kube-api-access-8qshn\") on node \"crc\" DevicePath \"\"" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.446412 4725 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e4c2870-ab09-4794-b3c9-836df98015ad-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.701576 4725 generic.go:334] "Generic (PLEG): container finished" podID="4e4c2870-ab09-4794-b3c9-836df98015ad" containerID="a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602" exitCode=0 Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.701624 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pllrb" event={"ID":"4e4c2870-ab09-4794-b3c9-836df98015ad","Type":"ContainerDied","Data":"a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602"} Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.701654 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pllrb" event={"ID":"4e4c2870-ab09-4794-b3c9-836df98015ad","Type":"ContainerDied","Data":"3c2abdf367e1fbe34a33e1dd4089ec679ff9e7f336e8b789a970e8d2ff7ff6b6"} Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.701679 4725 scope.go:117] "RemoveContainer" containerID="a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.701699 4725 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pllrb" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.721177 4725 scope.go:117] "RemoveContainer" containerID="3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.763360 4725 scope.go:117] "RemoveContainer" containerID="1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.763600 4725 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pllrb"] Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.775048 4725 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pllrb"] Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.789723 4725 scope.go:117] "RemoveContainer" containerID="a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602" Dec 02 14:16:12 crc kubenswrapper[4725]: E1202 14:16:12.790683 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602\": container with ID starting with a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602 not found: ID does not exist" containerID="a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.790737 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602"} err="failed to get container status \"a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602\": rpc error: code = NotFound desc = could not find container \"a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602\": container with ID starting with a223f5e9e7263c255c5a5574c2cbf3e0ef86c85f7af0f100e9efe68782e84602 not found: ID does not exist" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.790766 4725 scope.go:117] "RemoveContainer" containerID="3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d" Dec 02 14:16:12 crc kubenswrapper[4725]: E1202 14:16:12.791092 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d\": container with ID starting with 3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d not found: ID does not exist" containerID="3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.791128 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d"} err="failed to get container status \"3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d\": rpc error: code = NotFound desc = could not find container \"3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d\": container with ID starting with 3dbfb44aac99dc3250f00f4b8ddf25c93214dd66350396cdff4758412cfb9c5d not found: ID does not exist" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.791151 4725 scope.go:117] "RemoveContainer" containerID="1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e" Dec 02 14:16:12 crc kubenswrapper[4725]: E1202 14:16:12.791428 4725 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e\": container with ID starting with 1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e not found: ID does not exist" containerID="1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e" Dec 02 14:16:12 crc kubenswrapper[4725]: I1202 14:16:12.791448 4725 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e"} err="failed to get container status \"1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e\": rpc error: code = NotFound desc = could not find container \"1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e\": container with ID starting with 1c4dacaecff91967f91003ddc648e1394b959ef6971319d9ebfcc8fe1ef8236e not found: ID does not exist" Dec 02 14:16:13 crc kubenswrapper[4725]: I1202 14:16:13.289099 4725 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e4c2870-ab09-4794-b3c9-836df98015ad" path="/var/lib/kubelet/pods/4e4c2870-ab09-4794-b3c9-836df98015ad/volumes" Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.284596 4725 patch_prober.go:28] interesting pod/machine-config-daemon-zc2vm container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.285128 4725 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.285165 4725 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.285774 4725 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438"} pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.285826 4725 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerName="machine-config-daemon" containerID="cri-o://669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438" gracePeriod=600 Dec 02 14:16:24 crc kubenswrapper[4725]: E1202 14:16:24.411276 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.824207 4725 generic.go:334] "Generic (PLEG): container finished" podID="c50af795-4a0f-45cd-b117-f6ddea79ee51" containerID="669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438" exitCode=0 Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.824279 4725 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" event={"ID":"c50af795-4a0f-45cd-b117-f6ddea79ee51","Type":"ContainerDied","Data":"669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438"} Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.824342 4725 scope.go:117] "RemoveContainer" containerID="5996f0084ac41bbc5daaf0acbce3974adee0d79875ac90bac80ddb871f615a6d" Dec 02 14:16:24 crc kubenswrapper[4725]: I1202 14:16:24.825319 4725 scope.go:117] "RemoveContainer" containerID="669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438" Dec 02 14:16:24 crc kubenswrapper[4725]: E1202 14:16:24.825911 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:16:37 crc kubenswrapper[4725]: I1202 14:16:37.274346 4725 scope.go:117] "RemoveContainer" containerID="669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438" Dec 02 14:16:37 crc kubenswrapper[4725]: E1202 14:16:37.275112 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:16:50 crc kubenswrapper[4725]: I1202 14:16:50.268695 4725 scope.go:117] "RemoveContainer" containerID="669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438" Dec 02 14:16:50 crc kubenswrapper[4725]: E1202 14:16:50.269387 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" Dec 02 14:17:01 crc kubenswrapper[4725]: I1202 14:17:01.268014 4725 scope.go:117] "RemoveContainer" containerID="669cdfad3758f6cfe662b2600046ee25b1792a083c3e9fa7af7bbf571bf33438" Dec 02 14:17:01 crc kubenswrapper[4725]: E1202 14:17:01.268681 4725 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zc2vm_openshift-machine-config-operator(c50af795-4a0f-45cd-b117-f6ddea79ee51)\"" pod="openshift-machine-config-operator/machine-config-daemon-zc2vm" podUID="c50af795-4a0f-45cd-b117-f6ddea79ee51" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113572346024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113572346017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113561443016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113561443015460 5ustar corecore